/Linux-v5.4/drivers/scsi/megaraid/ |
D | megaraid_sas_fp.c | 124 u16 MR_LdSpanArrayGet(u32 ld, u32 span, struct MR_DRV_RAID_MAP_ALL *map) in MR_LdSpanArrayGet() argument 126 return le16_to_cpu(map->raidMap.ldSpanMap[ld].spanBlock[span].span.arrayRef); in MR_LdSpanArrayGet() 149 static struct MR_LD_SPAN *MR_LdSpanPtrGet(u32 ld, u32 span, in MR_LdSpanPtrGet() argument 152 return &map->raidMap.ldSpanMap[ld].spanBlock[span].span; in MR_LdSpanPtrGet() 376 u32 span, j; in MR_GetSpanBlock() local 378 for (span = 0; span < raid->spanDepth; span++, pSpanBlock++) { in MR_GetSpanBlock() 396 return span; in MR_GetSpanBlock() 428 u32 span, info; in mr_spanset_get_span_block() local 440 for (span = 0; span < raid->spanDepth; span++) in mr_spanset_get_span_block() 441 if (le32_to_cpu(map->raidMap.ldSpanMap[ld].spanBlock[span]. in mr_spanset_get_span_block() [all …]
|
D | mbox_defs.h | 649 adap_span_40ld_t span[SPAN_DEPTH_8_SPANS]; member 662 adap_span_8ld_t span[SPAN_DEPTH_8_SPANS]; member 675 adap_span_8ld_t span[SPAN_DEPTH_4_SPANS]; member
|
/Linux-v5.4/kernel/power/ |
D | energy_model.c | 81 static struct em_perf_domain *em_create_pd(cpumask_t *span, int nr_states, in em_create_pd() argument 86 int i, ret, cpu = cpumask_first(span); in em_create_pd() 158 cpumask_copy(to_cpumask(pd->cpus), span); in em_create_pd() 199 int em_register_perf_domain(cpumask_t *span, unsigned int nr_states, in em_register_perf_domain() argument 206 if (!span || !nr_states || !cb) in em_register_perf_domain() 215 for_each_cpu(cpu, span) { in em_register_perf_domain() 229 cpumask_pr_args(span)); in em_register_perf_domain() 237 pd = em_create_pd(span, nr_states, cb); in em_register_perf_domain() 243 for_each_cpu(cpu, span) { in em_register_perf_domain() 252 pr_debug("Created perf domain %*pbl\n", cpumask_pr_args(span)); in em_register_perf_domain()
|
/Linux-v5.4/drivers/net/ethernet/mellanox/mlxsw/ |
D | spectrum_span.c | 24 mlxsw_sp->span.entries_count = MLXSW_CORE_RES_GET(mlxsw_sp->core, in mlxsw_sp_span_init() 26 mlxsw_sp->span.entries = kcalloc(mlxsw_sp->span.entries_count, in mlxsw_sp_span_init() 29 if (!mlxsw_sp->span.entries) in mlxsw_sp_span_init() 32 for (i = 0; i < mlxsw_sp->span.entries_count; i++) { in mlxsw_sp_span_init() 33 struct mlxsw_sp_span_entry *curr = &mlxsw_sp->span.entries[i]; in mlxsw_sp_span_init() 46 for (i = 0; i < mlxsw_sp->span.entries_count; i++) { in mlxsw_sp_span_fini() 47 struct mlxsw_sp_span_entry *curr = &mlxsw_sp->span.entries[i]; in mlxsw_sp_span_fini() 51 kfree(mlxsw_sp->span.entries); in mlxsw_sp_span_fini() 627 for (i = 0; i < mlxsw_sp->span.entries_count; i++) { in mlxsw_sp_span_entry_create() 628 if (!mlxsw_sp->span.entries[i].ref_count) { in mlxsw_sp_span_entry_create() [all …]
|
/Linux-v5.4/Documentation/scheduler/ |
D | sched-domains.rst | 10 Each scheduling domain spans a number of CPUs (stored in the ->span field). 11 A domain's span MUST be a superset of it child's span (this restriction could 12 be relaxed if the need arises), and a base domain for CPU i MUST span at least 13 i. The top domain for each CPU will generally span all CPUs in the system 16 explicitly set. A sched domain's span means "balance process load among these 22 domain's span. The intersection of cpumasks from any two of these groups 56 The "base" domain will "span" the first level of the hierarchy. In the case 57 of SMT, you'll span all siblings of the physical CPU, with each group being 60 In SMP, the parent of the base domain will span all physical CPUs in the 62 of the SMP domain will span the entire machine, with each group having the
|
/Linux-v5.4/drivers/iio/dac/ |
D | ad5791.c | 68 int (*get_lin_comp) (unsigned int span); 212 static int ad5791_get_lin_comp(unsigned int span) in ad5791_get_lin_comp() argument 214 if (span <= 10000) in ad5791_get_lin_comp() 216 else if (span <= 12000) in ad5791_get_lin_comp() 218 else if (span <= 16000) in ad5791_get_lin_comp() 220 else if (span <= 19000) in ad5791_get_lin_comp() 226 static int ad5780_get_lin_comp(unsigned int span) in ad5780_get_lin_comp() argument 228 if (span <= 10000) in ad5780_get_lin_comp()
|
/Linux-v5.4/arch/sh/mm/ |
D | pmb.c | 146 unsigned long span; in pmb_mapping_exists() local 169 span = pmbe->size; in pmb_mapping_exists() 176 span += iter->size; in pmb_mapping_exists() 181 if (size <= span) { in pmb_mapping_exists() 657 unsigned long span, newsize; in pmb_merge() local 661 span = newsize = head->size; in pmb_merge() 665 span += tail->size; in pmb_merge() 667 if (pmb_size_valid(span)) { in pmb_merge() 668 newsize = span; in pmb_merge()
|
/Linux-v5.4/include/linux/ |
D | energy_model.h | 67 int em_register_perf_domain(cpumask_t *span, unsigned int nr_states, 167 static inline int em_register_perf_domain(cpumask_t *span, in em_register_perf_domain() argument
|
/Linux-v5.4/net/mac802154/ |
D | rx.c | 37 __le16 span, sshort; in ieee802154_subif_frame() local 42 span = wpan_dev->pan_id; in ieee802154_subif_frame() 55 if (mac_cb(skb)->dest.pan_id != span && in ieee802154_subif_frame() 64 if (mac_cb(skb)->dest.pan_id != span && in ieee802154_subif_frame()
|
/Linux-v5.4/include/linux/sched/ |
D | topology.h | 145 unsigned long span[0]; member 150 return to_cpumask(sd->span); in sched_domain_span()
|
/Linux-v5.4/drivers/perf/ |
D | arm_smmuv3_pmu.c | 251 int idx, u32 span, u32 sid) in smmu_pmu_set_event_filter() argument 256 evtyper = get_event(event) | span << SMMU_PMCG_SID_SPAN_SHIFT; in smmu_pmu_set_event_filter() 277 u32 span, sid; in smmu_pmu_apply_event_filter() local 281 span = filter_en ? get_filter_span(event) : in smmu_pmu_apply_event_filter() 288 smmu_pmu_set_event_filter(event, idx, span, sid); in smmu_pmu_apply_event_filter() 296 smmu_pmu_set_event_filter(event, 0, span, sid); in smmu_pmu_apply_event_filter()
|
/Linux-v5.4/Documentation/sphinx-static/ |
D | theme_overrides.css | 96 span.menuselection { 101 code.kbd, code.kbd span {
|
/Linux-v5.4/drivers/s390/scsi/ |
D | zfcp_qdio.c | 52 unsigned long long now, span; in zfcp_qdio_account() local 56 span = (now - qdio->req_q_time) >> 12; in zfcp_qdio_account() 58 qdio->req_q_util += used * span; in zfcp_qdio_account()
|
/Linux-v5.4/drivers/usb/host/ |
D | ehci-sched.c | 1204 iso_sched->span = urb->number_of_packets * stream->uperiod; in itd_sched_init() 1269 num_itds = 1 + (sched->span + 7) / 8; in itd_urb_transaction() 1486 u32 now, base, next, start, period, span, now2; in iso_stream_schedule() local 1495 span = sched->span; in iso_stream_schedule() 1497 span <<= 3; in iso_stream_schedule() 1622 if (skip >= span) { /* Entirely in the past? */ in iso_stream_schedule() 1624 urb, start + base, span - period, now2 + base, in iso_stream_schedule() 1628 skip = span - period; in iso_stream_schedule() 1632 skip = span; /* Skip the entire URB */ in iso_stream_schedule() 1649 if (unlikely(start + span - period >= mod + wrap)) { in iso_stream_schedule() [all …]
|
/Linux-v5.4/kernel/sched/ |
D | topology.c | 435 free_cpumask_var(rd->span); in free_rootdomain() 453 cpumask_clear_cpu(rq->cpu, old_rd->span); in rq_attach_root() 467 cpumask_set_cpu(rq->cpu, rd->span); in rq_attach_root() 492 if (!zalloc_cpumask_var(&rd->span, GFP_KERNEL)) in init_rootdomain() 524 free_cpumask_var(rd->span); in init_rootdomain() 930 const struct cpumask *span = sched_domain_span(sd); in build_overlap_sched_groups() local 938 for_each_cpu_wrap(i, span, cpu) { in build_overlap_sched_groups() 1106 const struct cpumask *span = sched_domain_span(sd); in build_sched_groups() local 1115 for_each_cpu_wrap(i, span, cpu) { in build_sched_groups()
|
D | rt.c | 543 return this_rq()->rd->span; in sched_rt_period_mask() 646 weight = cpumask_weight(rd->span); in do_balance_runtime() 650 for_each_cpu(i, rd->span) { in do_balance_runtime() 727 for_each_cpu(i, rd->span) { in __disable_runtime() 815 const struct cpumask *span; in do_sched_rt_period_timer() local 817 span = sched_rt_period_mask(); in do_sched_rt_period_timer() 829 span = cpu_online_mask; in do_sched_rt_period_timer() 831 for_each_cpu(i, span) { in do_sched_rt_period_timer()
|
/Linux-v5.4/Documentation/sphinx/ |
D | rstFlatTable.py | 87 nodelist = [colSpan(span=int(text))] 98 nodelist = [rowSpan(span=int(text))]
|
/Linux-v5.4/mm/ |
D | page_io.c | 148 sector_t *span) in generic_swapfile_activate() argument 226 *span = 1 + highest_block - lowest_block; in generic_swapfile_activate()
|
D | swapfile.c | 2381 static int setup_swap_extents(struct swap_info_struct *sis, sector_t *span) in setup_swap_extents() argument 2390 *span = sis->pages; in setup_swap_extents() 2395 ret = mapping->a_ops->swap_activate(sis, swap_file, span); in setup_swap_extents() 2401 *span = sis->pages; in setup_swap_extents() 2406 return generic_swapfile_activate(sis, swap_file, span); in setup_swap_extents() 3012 sector_t *span) in setup_swap_map_and_extents() argument 3054 nr_extents = setup_swap_extents(p, span); in setup_swap_map_and_extents() 3112 sector_t span; in SYSCALL_DEFINE2() local 3233 cluster_info, maxpages, &span); in SYSCALL_DEFINE2() 3298 nr_extents, (unsigned long long)span<<(PAGE_SHIFT-10), in SYSCALL_DEFINE2()
|
/Linux-v5.4/tools/perf/ui/stdio/ |
D | hist.c | 711 int span = 0; in fprintf_line() local 717 if (!first && !span) in fprintf_line() 722 fmt->header(fmt, hpp, hists, line, &span); in fprintf_line() 724 if (!span) in fprintf_line()
|
/Linux-v5.4/Documentation/filesystems/ext4/ |
D | inlinedata.rst | 34 entries cannot span the two.
|
/Linux-v5.4/Documentation/doc-guide/ |
D | sphinx.rst | 267 * column-span: with the role ``cspan`` a cell can be extended through 270 * row-span: with the role ``rspan`` a cell can be extended through 273 * auto span rightmost cell of a table row over the missing cells on the right 275 changed from *auto span* to *auto fill*, which automatically inserts (empty)
|
/Linux-v5.4/Documentation/translations/it_IT/doc-guide/ |
D | sphinx.rst | 293 * column-span: col ruolo ``cspan`` una cella può essere estesa attraverso 296 * raw-span: col ruolo ``rspan`` una cella può essere estesa attraverso 299 * auto-span: la cella più a destra viene estesa verso destra per compensare 301 può essere cambiato da *auto-span* ad *auto-fill*, il quale inserisce
|
/Linux-v5.4/drivers/irqchip/ |
D | irq-gic-v3-its.c | 1477 u32 span; member 1480 static struct lpi_range *mk_lpi_range(u32 base, u32 span) in mk_lpi_range() argument 1487 range->span = span; in mk_lpi_range() 1501 if (range->span >= nr_lpis) { in alloc_lpi_range() 1504 range->span -= nr_lpis; in alloc_lpi_range() 1506 if (range->span == 0) { in alloc_lpi_range() 1526 if (a->base_id + a->span != b->base_id) in merge_lpi_ranges() 1529 b->span += a->span; in merge_lpi_ranges()
|
/Linux-v5.4/Documentation/vm/ |
D | zsmalloc.rst | 17 pages act as a single higher-order page i.e. an object can span 0-order
|