/Linux-v5.4/kernel/events/ |
D | callchain.c | 50 struct callchain_cpus_entries *entries; in release_callchain_buffers_rcu() local 53 entries = container_of(head, struct callchain_cpus_entries, rcu_head); in release_callchain_buffers_rcu() 56 kfree(entries->cpu_entries[cpu]); in release_callchain_buffers_rcu() 58 kfree(entries); in release_callchain_buffers_rcu() 63 struct callchain_cpus_entries *entries; in release_callchain_buffers() local 65 entries = callchain_cpus_entries; in release_callchain_buffers() 67 call_rcu(&entries->rcu_head, release_callchain_buffers_rcu); in release_callchain_buffers() 74 struct callchain_cpus_entries *entries; in alloc_callchain_buffers() local 83 entries = kzalloc(size, GFP_KERNEL); in alloc_callchain_buffers() 84 if (!entries) in alloc_callchain_buffers() [all …]
|
/Linux-v5.4/drivers/gpu/drm/amd/powerplay/hwmgr/ |
D | smu_helper.c | 223 vvalue = vol_table->entries[i].value; in phm_trim_voltage_table() 227 if (vvalue == table->entries[j].value) { in phm_trim_voltage_table() 234 table->entries[table->count].value = vvalue; in phm_trim_voltage_table() 235 table->entries[table->count].smio_low = in phm_trim_voltage_table() 236 vol_table->entries[i].smio_low; in phm_trim_voltage_table() 264 vol_table->entries[i].value = dep_table->entries[i].mvdd; in phm_get_svi2_mvdd_voltage_table() 265 vol_table->entries[i].smio_low = 0; in phm_get_svi2_mvdd_voltage_table() 292 vol_table->entries[i].value = dep_table->entries[i].vddci; in phm_get_svi2_vddci_voltage_table() 293 vol_table->entries[i].smio_low = 0; in phm_get_svi2_vddci_voltage_table() 320 vol_table->entries[i].value = lookup_table->entries[i].us_vdd; in phm_get_svi2_vdd_voltage_table() [all …]
|
D | vega10_processpptables.c | 320 (ATOM_Vega10_GFXCLK_Dependency_Record_V2 *)gfxclk_dep_table->entries; in init_over_drive_limits() 365 mm_dependency_record = &mm_dependency_table->entries[i]; in get_mm_clock_voltage_table() 366 mm_table->entries[i].vddcInd = mm_dependency_record->ucVddcInd; in get_mm_clock_voltage_table() 367 mm_table->entries[i].samclock = in get_mm_clock_voltage_table() 369 mm_table->entries[i].eclk = le32_to_cpu(mm_dependency_record->ulEClk); in get_mm_clock_voltage_table() 370 mm_table->entries[i].vclk = le32_to_cpu(mm_dependency_record->ulVClk); in get_mm_clock_voltage_table() 371 mm_table->entries[i].dclk = le32_to_cpu(mm_dependency_record->ulDClk); in get_mm_clock_voltage_table() 592 clk_table->entries[i].vddInd = in get_socclk_voltage_dependency_table() 593 clk_dep_table->entries[i].ucVddInd; in get_socclk_voltage_dependency_table() 594 clk_table->entries[i].clk = in get_socclk_voltage_dependency_table() [all …]
|
/Linux-v5.4/arch/powerpc/mm/book3s64/ |
D | iommu_api.c | 34 u64 entries; /* number of entries in hpas/hpages[] */ member 57 unsigned long entries, unsigned long dev_hpa, in mm_iommu_do_alloc() argument 66 ret = account_locked_vm(mm, entries, true); in mm_iommu_do_alloc() 70 locked_entries = entries; in mm_iommu_do_alloc() 80 mem->pageshift = __ffs(dev_hpa | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc() 91 mem->pageshift = __ffs(ua | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc() 92 mem->hpas = vzalloc(array_size(entries, sizeof(mem->hpas[0]))); in mm_iommu_do_alloc() 102 chunk = min(chunk, entries); in mm_iommu_do_alloc() 103 for (entry = 0; entry < entries; entry += chunk) { in mm_iommu_do_alloc() 104 unsigned long n = min(entries - entry, chunk); in mm_iommu_do_alloc() [all …]
|
/Linux-v5.4/tools/perf/util/ |
D | mem2node.c | 49 struct phys_entry *entries, *tmp_entries; in mem2node__init() local 61 entries = zalloc(sizeof(*entries) * max); in mem2node__init() 62 if (!entries) in mem2node__init() 83 struct phys_entry *prev = &entries[j - 1]; in mem2node__init() 92 phys_entry__init(&entries[j++], start, bsize, n->node); in mem2node__init() 97 tmp_entries = realloc(entries, sizeof(*entries) * j); in mem2node__init() 99 entries = tmp_entries; in mem2node__init() 103 entries[i].node, entries[i].start, entries[i].end); in mem2node__init() 105 phys_entry__insert(&entries[i], &map->root); in mem2node__init() 108 map->entries = entries; in mem2node__init() [all …]
|
D | rb_resort.h | 72 struct rb_root entries; \ 79 struct rb_node **p = &sorted->entries.rb_node, *parent = NULL; \ 88 rb_insert_color(sorted_nd, &sorted->entries); \ 92 struct rb_root *entries) \ 96 for (nd = rb_first(entries); nd; nd = rb_next(nd)) { \ 103 static struct __name##_sorted *__name##_sorted__new(struct rb_root *entries, \ 109 sorted->entries = RB_ROOT; \ 110 __name##_sorted__sort(sorted, entries); \ 128 for (__nd = rb_first(&__name->entries); \ 143 DECLARE_RESORT_RB(__name)(&__ilist->rblist.entries.rb_root, \ [all …]
|
D | pstack.c | 18 void *entries[0]; member 45 if (pstack->entries[i] == key) { in pstack__remove() 47 memmove(pstack->entries + i, in pstack__remove() 48 pstack->entries + i + 1, in pstack__remove() 63 pstack->entries[pstack->top++] = key; in pstack__push() 75 ret = pstack->entries[--pstack->top]; in pstack__pop() 76 pstack->entries[pstack->top] = NULL; in pstack__pop() 84 return pstack->entries[pstack->top - 1]; in pstack__peek()
|
D | syscalltbl.c | 62 struct syscall *entries; in syscalltbl__init_native() local 68 entries = tbl->syscalls.entries = malloc(sizeof(struct syscall) * nr_entries); in syscalltbl__init_native() 69 if (tbl->syscalls.entries == NULL) in syscalltbl__init_native() 74 entries[j].name = syscalltbl_native[i]; in syscalltbl__init_native() 75 entries[j].id = i; in syscalltbl__init_native() 80 qsort(tbl->syscalls.entries, nr_entries, sizeof(struct syscall), syscallcmp); in syscalltbl__init_native() 100 zfree(&tbl->syscalls.entries); in syscalltbl__delete() 111 struct syscall *sc = bsearch(name, tbl->syscalls.entries, in syscalltbl__id() 121 struct syscall *syscalls = tbl->syscalls.entries; in syscalltbl__strglobmatch_next()
|
/Linux-v5.4/tools/lib/api/fd/ |
D | array.c | 14 fda->entries = NULL; in fdarray__init() 26 struct pollfd *entries = realloc(fda->entries, size); in fdarray__grow() local 28 if (entries == NULL) in fdarray__grow() 33 free(entries); in fdarray__grow() 38 fda->entries = entries; in fdarray__grow() 61 free(fda->entries); in fdarray__exit() 80 fda->entries[fda->nr].fd = fd; in fdarray__add() 81 fda->entries[fda->nr].events = revents; in fdarray__add() 96 if (fda->entries[fd].revents & revents) { in fdarray__filter() 104 fda->entries[nr] = fda->entries[fd]; in fdarray__filter() [all …]
|
/Linux-v5.4/arch/x86/kernel/cpu/ |
D | intel.c | 871 if (tlb_lli_4k[ENTRIES] < intel_tlb_table[k].entries) in intel_tlb_lookup() 872 tlb_lli_4k[ENTRIES] = intel_tlb_table[k].entries; in intel_tlb_lookup() 873 if (tlb_lld_4k[ENTRIES] < intel_tlb_table[k].entries) in intel_tlb_lookup() 874 tlb_lld_4k[ENTRIES] = intel_tlb_table[k].entries; in intel_tlb_lookup() 877 if (tlb_lli_4k[ENTRIES] < intel_tlb_table[k].entries) in intel_tlb_lookup() 878 tlb_lli_4k[ENTRIES] = intel_tlb_table[k].entries; in intel_tlb_lookup() 879 if (tlb_lld_4k[ENTRIES] < intel_tlb_table[k].entries) in intel_tlb_lookup() 880 tlb_lld_4k[ENTRIES] = intel_tlb_table[k].entries; in intel_tlb_lookup() 881 if (tlb_lli_2m[ENTRIES] < intel_tlb_table[k].entries) in intel_tlb_lookup() 882 tlb_lli_2m[ENTRIES] = intel_tlb_table[k].entries; in intel_tlb_lookup() [all …]
|
/Linux-v5.4/drivers/net/ethernet/netronome/nfp/nfpcore/ |
D | nfp_nsp_eth.c | 247 union eth_table_entry *entries; in __nfp_eth_read_ports() local 251 entries = kzalloc(NSP_ETH_TABLE_SIZE, GFP_KERNEL); in __nfp_eth_read_ports() 252 if (!entries) in __nfp_eth_read_ports() 255 ret = nfp_nsp_read_eth_table(nsp, entries, NSP_ETH_TABLE_SIZE); in __nfp_eth_read_ports() 262 if (entries[i].port & NSP_ETH_PORT_LANES_MASK) in __nfp_eth_read_ports() 281 if (entries[i].port & NSP_ETH_PORT_LANES_MASK) in __nfp_eth_read_ports() 282 nfp_eth_port_translate(nsp, &entries[i], i, in __nfp_eth_read_ports() 289 kfree(entries); in __nfp_eth_read_ports() 294 kfree(entries); in __nfp_eth_read_ports() 300 union eth_table_entry *entries; in nfp_eth_config_start() local [all …]
|
/Linux-v5.4/tools/perf/tests/ |
D | fdarray.c | 14 fda->entries[fd].fd = fda->nr - fd; in fdarray__init_revents() 15 fda->entries[fd].revents = revents; in fdarray__init_revents() 57 fda->entries[2].revents = POLLIN; in test__fdarray__filter() 58 expected_fd[0] = fda->entries[2].fd; in test__fdarray__filter() 69 if (fda->entries[0].fd != expected_fd[0]) { in test__fdarray__filter() 71 fda->entries[0].fd, expected_fd[0]); in test__fdarray__filter() 76 fda->entries[0].revents = POLLIN; in test__fdarray__filter() 77 expected_fd[0] = fda->entries[0].fd; in test__fdarray__filter() 78 fda->entries[3].revents = POLLIN; in test__fdarray__filter() 79 expected_fd[1] = fda->entries[3].fd; in test__fdarray__filter() [all …]
|
/Linux-v5.4/lib/ |
D | stackdepot.c | 64 unsigned long entries[1]; /* Variable-sized array of entries. */ member 99 static struct stack_record *depot_alloc_stack(unsigned long *entries, int size, in depot_alloc_stack() argument 102 int required_size = offsetof(struct stack_record, entries) + in depot_alloc_stack() 134 memcpy(stack->entries, entries, size * sizeof(unsigned long)); in depot_alloc_stack() 150 static inline u32 hash_stack(unsigned long *entries, unsigned int size) in hash_stack() argument 152 return jhash2((u32 *)entries, in hash_stack() 174 unsigned long *entries, int size, in find_stack() argument 182 !stackdepot_memcmp(entries, found->entries, size)) in find_stack() 198 unsigned long **entries) in stack_depot_fetch() argument 205 *entries = stack->entries; in stack_depot_fetch() [all …]
|
D | test_rhashtable.c | 73 unsigned int entries; member 138 unsigned int entries) in test_rht_lookup() argument 142 for (i = 0; i < entries; i++) { in test_rht_lookup() 175 static void test_bucket_stats(struct rhashtable *ht, unsigned int entries) in test_bucket_stats() argument 202 total, atomic_read(&ht->nelems), entries, chain_len); in test_bucket_stats() 204 if (total != atomic_read(&ht->nelems) || total != entries) in test_bucket_stats() 209 unsigned int entries) in test_rhashtable() argument 220 pr_info(" Adding %d keys\n", entries); in test_rhashtable() 222 for (i = 0; i < entries; i++) { in test_rhashtable() 237 test_bucket_stats(ht, entries); in test_rhashtable() [all …]
|
/Linux-v5.4/drivers/net/dsa/sja1105/ |
D | sja1105_tas.c | 107 kfree(table->entries); in sja1105_init_scheduling() 114 kfree(table->entries); in sja1105_init_scheduling() 121 kfree(table->entries); in sja1105_init_scheduling() 128 kfree(table->entries); in sja1105_init_scheduling() 148 table->entries = kcalloc(num_entries, table->ops->unpacked_entry_size, in sja1105_init_scheduling() 150 if (!table->entries) in sja1105_init_scheduling() 153 schedule = table->entries; in sja1105_init_scheduling() 157 table->entries = kcalloc(SJA1105_MAX_SCHEDULE_ENTRY_POINTS_PARAMS_COUNT, in sja1105_init_scheduling() 159 if (!table->entries) in sja1105_init_scheduling() 166 schedule_entry_points_params = table->entries; in sja1105_init_scheduling() [all …]
|
/Linux-v5.4/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_dpm.c | 234 amdgpu_table->entries = kzalloc(size, GFP_KERNEL); in amdgpu_parse_clk_voltage_dep_table() 235 if (!amdgpu_table->entries) in amdgpu_parse_clk_voltage_dep_table() 238 entry = &atom_table->entries[0]; in amdgpu_parse_clk_voltage_dep_table() 240 amdgpu_table->entries[i].clk = le16_to_cpu(entry->usClockLow) | in amdgpu_parse_clk_voltage_dep_table() 242 amdgpu_table->entries[i].v = le16_to_cpu(entry->usVoltage); in amdgpu_parse_clk_voltage_dep_table() 381 le16_to_cpu(clk_v->entries[0].usSclkLow) | in amdgpu_parse_extended_power_table() 382 (clk_v->entries[0].ucSclkHigh << 16); in amdgpu_parse_extended_power_table() 384 le16_to_cpu(clk_v->entries[0].usMclkLow) | in amdgpu_parse_extended_power_table() 385 (clk_v->entries[0].ucMclkHigh << 16); in amdgpu_parse_extended_power_table() 387 le16_to_cpu(clk_v->entries[0].usVddc); in amdgpu_parse_extended_power_table() [all …]
|
/Linux-v5.4/tools/perf/trace/beauty/ |
D | ioctl.c | 41 if (nr < strarray__ioctl_tty_cmd.nr_entries && strarray__ioctl_tty_cmd.entries[nr] != NULL) in ioctl__scnprintf_tty_cmd() 42 return scnprintf(bf, size, "%s", strarray__ioctl_tty_cmd.entries[nr]); in ioctl__scnprintf_tty_cmd() 52 if (nr < strarray__drm_ioctl_cmds.nr_entries && strarray__drm_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_drm_cmd() 53 return scnprintf(bf, size, "DRM_%s", strarray__drm_ioctl_cmds.entries[nr]); in ioctl__scnprintf_drm_cmd() 63 …if (nr < strarray__sndrv_pcm_ioctl_cmds.nr_entries && strarray__sndrv_pcm_ioctl_cmds.entries[nr] !… in ioctl__scnprintf_sndrv_pcm_cmd() 64 return scnprintf(bf, size, "SNDRV_PCM_%s", strarray__sndrv_pcm_ioctl_cmds.entries[nr]); in ioctl__scnprintf_sndrv_pcm_cmd() 74 …if (nr < strarray__sndrv_ctl_ioctl_cmds.nr_entries && strarray__sndrv_ctl_ioctl_cmds.entries[nr] !… in ioctl__scnprintf_sndrv_ctl_cmd() 75 return scnprintf(bf, size, "SNDRV_CTL_%s", strarray__sndrv_ctl_ioctl_cmds.entries[nr]); in ioctl__scnprintf_sndrv_ctl_cmd() 85 if (nr < strarray__kvm_ioctl_cmds.nr_entries && strarray__kvm_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_kvm_cmd() 86 return scnprintf(bf, size, "KVM_%s", strarray__kvm_ioctl_cmds.entries[nr]); in ioctl__scnprintf_kvm_cmd() [all …]
|
/Linux-v5.4/drivers/misc/vmw_vmci/ |
D | vmci_handle_array.c | 68 array->entries[array->size] = handle; in vmci_handle_arr_append_entry() 84 if (vmci_handle_is_equal(array->entries[i], entry_handle)) { in vmci_handle_arr_remove_entry() 85 handle = array->entries[i]; in vmci_handle_arr_remove_entry() 87 array->entries[i] = array->entries[array->size]; in vmci_handle_arr_remove_entry() 88 array->entries[array->size] = VMCI_INVALID_HANDLE; in vmci_handle_arr_remove_entry() 105 handle = array->entries[array->size]; in vmci_handle_arr_remove_tail() 106 array->entries[array->size] = VMCI_INVALID_HANDLE; in vmci_handle_arr_remove_tail() 121 return array->entries[index]; in vmci_handle_arr_get_entry() 130 if (vmci_handle_is_equal(array->entries[i], entry_handle)) in vmci_handle_arr_has_entry() 143 return array->entries; in vmci_handle_arr_get_handles()
|
/Linux-v5.4/arch/x86/kernel/ |
D | e820.c | 83 struct e820_entry *entry = &table->entries[i]; in _e820__mapped_any() 118 struct e820_entry *entry = &e820_table->entries[i]; in __e820__mapped_all() 170 if (x >= ARRAY_SIZE(table->entries)) { in __e820__range_add() 176 table->entries[x].addr = start; in __e820__range_add() 177 table->entries[x].size = size; in __e820__range_add() 178 table->entries[x].type = type; in __e820__range_add() 209 e820_table->entries[i].addr, in e820__print_table() 210 e820_table->entries[i].addr + e820_table->entries[i].size - 1); in e820__print_table() 212 e820_print_type(e820_table->entries[i].type); in e820__print_table() 309 struct e820_entry *entries = table->entries; in e820__update_table() local [all …]
|
/Linux-v5.4/arch/x86/xen/ |
D | multicalls.c | 41 struct multicall_entry entries[MC_BATCH]; member 73 memcpy(b->debug, b->entries, in xen_mc_flush() 86 mc = &b->entries[0]; in xen_mc_flush() 95 if (HYPERVISOR_multicall(b->entries, b->mcidx) != 0) in xen_mc_flush() 98 if (b->entries[i].result < 0) in xen_mc_flush() 106 if (b->entries[i].result < 0) { in xen_mc_flush() 112 b->entries[i].result, in xen_mc_flush() 117 b->entries[i].op, in xen_mc_flush() 118 b->entries[i].args[0], in xen_mc_flush() 119 b->entries[i].result); in xen_mc_flush() [all …]
|
/Linux-v5.4/virt/kvm/arm/vgic/ |
D | vgic-irqfd.c | 123 struct kvm_irq_routing_entry *entries; in kvm_vgic_setup_default_irq_routing() local 128 entries = kcalloc(nr, sizeof(*entries), GFP_KERNEL); in kvm_vgic_setup_default_irq_routing() 129 if (!entries) in kvm_vgic_setup_default_irq_routing() 133 entries[i].gsi = i; in kvm_vgic_setup_default_irq_routing() 134 entries[i].type = KVM_IRQ_ROUTING_IRQCHIP; in kvm_vgic_setup_default_irq_routing() 135 entries[i].u.irqchip.irqchip = 0; in kvm_vgic_setup_default_irq_routing() 136 entries[i].u.irqchip.pin = i; in kvm_vgic_setup_default_irq_routing() 138 ret = kvm_set_irq_routing(kvm, entries, nr, 0); in kvm_vgic_setup_default_irq_routing() 139 kfree(entries); in kvm_vgic_setup_default_irq_routing()
|
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx4/ |
D | port.c | 68 table->entries[i] = 0; in mlx4_init_mac_table() 82 table->entries[i] = 0; in mlx4_init_vlan_table() 105 if (index < 0 || index >= table->max || !table->entries[index]) { in validate_index() 120 (MLX4_MAC_MASK & be64_to_cpu(table->entries[i]))) in find_index() 128 __be64 *entries) in mlx4_set_port_mac_table() argument 138 memcpy(mailbox->buf, entries, MLX4_MAC_TABLE_SIZE); in mlx4_set_port_mac_table() 160 if (mac == (MLX4_MAC_MASK & be64_to_cpu(table->entries[i]))) { in mlx4_find_cached_mac() 216 if (((MLX4_MAC_MASK & mac) == (MLX4_MAC_MASK & be64_to_cpu(table->entries[i])))) in __mlx4_register_mac() 218 if (((MLX4_MAC_MASK & mac) == (MLX4_MAC_MASK & be64_to_cpu(dup_table->entries[i])))) in __mlx4_register_mac() 243 ((MLX4_MAC_MASK & mac) == (MLX4_MAC_MASK & be64_to_cpu(table->entries[index_at_dup_port])))) in __mlx4_register_mac() [all …]
|
/Linux-v5.4/drivers/media/dvb-frontends/ |
D | dvb-pll.c | 66 } entries[]; member 79 .entries = { 102 .entries = { 125 .entries = { 143 .entries = { 158 .entries = { 180 .entries = { 194 .entries = { 220 .entries = { 253 .entries = { [all …]
|
/Linux-v5.4/drivers/gpu/drm/radeon/ |
D | r600_dpm.c | 828 radeon_table->entries = kzalloc(size, GFP_KERNEL); in r600_parse_clk_voltage_dep_table() 829 if (!radeon_table->entries) in r600_parse_clk_voltage_dep_table() 832 entry = &atom_table->entries[0]; in r600_parse_clk_voltage_dep_table() 834 radeon_table->entries[i].clk = le16_to_cpu(entry->usClockLow) | in r600_parse_clk_voltage_dep_table() 836 radeon_table->entries[i].v = le16_to_cpu(entry->usVoltage); in r600_parse_clk_voltage_dep_table() 938 kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries); in r600_parse_extended_power_table() 949 kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries); in r600_parse_extended_power_table() 950 kfree(rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk.entries); in r600_parse_extended_power_table() 961 kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries); in r600_parse_extended_power_table() 962 kfree(rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk.entries); in r600_parse_extended_power_table() [all …]
|
/Linux-v5.4/kernel/ |
D | stacktrace.c | 23 void stack_trace_print(const unsigned long *entries, unsigned int nr_entries, in stack_trace_print() argument 28 if (WARN_ON(!entries)) in stack_trace_print() 32 printk("%*c%pS\n", 1 + spaces, ' ', (void *)entries[i]); in stack_trace_print() 46 int stack_trace_snprint(char *buf, size_t size, const unsigned long *entries, in stack_trace_snprint() argument 51 if (WARN_ON(!entries)) in stack_trace_snprint() 56 (void *)entries[i]); in stack_trace_snprint() 276 .entries = store, in stack_trace_save() 300 .entries = store, in stack_trace_save_tsk() 323 .entries = store, in stack_trace_save_regs() 349 .entries = store, in stack_trace_save_tsk_reliable() [all …]
|