/Linux-v5.10/drivers/net/ethernet/mellanox/mlxsw/ |
D | spectrum_span.c | 46 struct mlxsw_sp_span *span; member 77 return atomic_read(&mlxsw_sp->span->active_entries_count); in mlxsw_sp_span_occ_get() 83 struct mlxsw_sp_span *span; in mlxsw_sp_span_init() local 90 span = kzalloc(struct_size(span, entries, entries_count), GFP_KERNEL); in mlxsw_sp_span_init() 91 if (!span) in mlxsw_sp_span_init() 93 refcount_set(&span->policer_id_base_ref_count, 0); in mlxsw_sp_span_init() 94 span->entries_count = entries_count; in mlxsw_sp_span_init() 95 atomic_set(&span->active_entries_count, 0); in mlxsw_sp_span_init() 96 mutex_init(&span->analyzed_ports_lock); in mlxsw_sp_span_init() 97 INIT_LIST_HEAD(&span->analyzed_ports_list); in mlxsw_sp_span_init() [all …]
|
/Linux-v5.10/drivers/scsi/megaraid/ |
D | megaraid_sas_fp.c | 123 u16 MR_LdSpanArrayGet(u32 ld, u32 span, struct MR_DRV_RAID_MAP_ALL *map) in MR_LdSpanArrayGet() argument 125 return le16_to_cpu(map->raidMap.ldSpanMap[ld].spanBlock[span].span.arrayRef); in MR_LdSpanArrayGet() 148 static struct MR_LD_SPAN *MR_LdSpanPtrGet(u32 ld, u32 span, in MR_LdSpanPtrGet() argument 151 return &map->raidMap.ldSpanMap[ld].spanBlock[span].span; in MR_LdSpanPtrGet() 338 dev_err(&instance->pdev->dev, "megasas: span map %x, pDrvRaidMap->totalSize : %x\n", in MR_ValidateMapInfo() 375 u32 span, j; in MR_GetSpanBlock() local 377 for (span = 0; span < raid->spanDepth; span++, pSpanBlock++) { in MR_GetSpanBlock() 394 return span; in MR_GetSpanBlock() 404 * This routine calculates the Span block for given row using spanset. 414 * span - Span number [all …]
|
D | mbox_defs.h | 595 * adap_span_40ld_t - 40LD span 607 * adap_span_8ld_t - 8LD span 645 * @span : span 649 adap_span_40ld_t span[SPAN_DEPTH_8_SPANS]; member 656 * @span : span 662 adap_span_8ld_t span[SPAN_DEPTH_8_SPANS]; member 669 * @span : span 675 adap_span_8ld_t span[SPAN_DEPTH_4_SPANS]; member
|
/Linux-v5.10/Documentation/scheduler/ |
D | sched-domains.rst | 10 Each scheduling domain spans a number of CPUs (stored in the ->span field). 11 A domain's span MUST be a superset of it child's span (this restriction could 12 be relaxed if the need arises), and a base domain for CPU i MUST span at least 13 i. The top domain for each CPU will generally span all CPUs in the system 16 explicitly set. A sched domain's span means "balance process load among these 22 domain's span. The group pointed to by the ->groups pointer MUST contain the CPU 58 The "base" domain will "span" the first level of the hierarchy. In the case 59 of SMT, you'll span all siblings of the physical CPU, with each group being 62 In SMP, the parent of the base domain will span all physical CPUs in the 64 of the SMP domain will span the entire machine, with each group having the
|
/Linux-v5.10/tools/testing/selftests/resctrl/ |
D | mbm_test.c | 18 show_bw_info(unsigned long *bw_imc, unsigned long *bw_resc, int span) in show_bw_info() argument 42 printf("# Span (MB): %d\n", span); in show_bw_info() 47 static int check_results(int span) in check_results() argument 79 show_bw_info(bw_imc, bw_resc, span); in check_results() 114 int mbm_bw_change(int span, int cpu_no, char *bw_report, char **benchmark_cmd) in mbm_bw_change() argument 120 .span = span, in mbm_bw_change() 138 ret = check_results(span); in mbm_bw_change()
|
D | cqm_test.c | 43 unsigned long span) in show_cache_info() argument 51 avg_diff = (long)abs(span - avg_llc_occu_resc); in show_cache_info() 53 diff_percent = (((float)span - avg_llc_occu_resc) / span) * 100; in show_cache_info() 69 printf("# llc_occu_exp (span): %lu\n", span); in show_cache_info() 104 show_cache_info(sum_llc_occu_resc, no_of_bits, param->span); in check_results() 155 .span = cache_size * n / count_of_bits, in cqm_resctrl_val() 161 sprintf(benchmark_cmd[1], "%lu", param.span); in cqm_resctrl_val()
|
D | resctrl.h | 45 * @span: Memory bytes accessed in each benchmark iteration 56 unsigned long span; member 87 int run_fill_buf(unsigned long span, int malloc_and_init_memory, int memflush, 90 int mbm_bw_change(int span, int cpu_no, char *bw_report, char **benchmark_cmd);
|
D | cat_test.c | 56 unsigned long span) in show_cache_info() argument 58 unsigned long allocated_cache_lines = span / 64; in show_cache_info() 111 show_cache_info(sum_llc_perf_miss, no_of_bits, param->span); in check_results() 179 param.span = cache_size * (count_of_bits - n) / count_of_bits; in cat_perf_miss_val() 200 param.span = cache_size * n / count_of_bits; in cat_perf_miss_val()
|
D | resctrl_tests.c | 60 int res, c, cpu_no = 1, span = 250, argc_new = argc, i, no_of_bits = 5; in main() local 147 sprintf(benchmark_cmd[1], "%d", span); in main() 165 res = mbm_bw_change(span, cpu_no, bw_report, benchmark_cmd); in main() 174 sprintf(benchmark_cmd[1], "%d", span); in main()
|
D | fill_buf.c | 193 int run_fill_buf(unsigned long span, int malloc_and_init_memory, in run_fill_buf() argument 196 unsigned long long cache_size = span; in run_fill_buf()
|
/Linux-v5.10/drivers/iio/dac/ |
D | ad5791.c | 68 int (*get_lin_comp) (unsigned int span); 210 static int ad5791_get_lin_comp(unsigned int span) in ad5791_get_lin_comp() argument 212 if (span <= 10000) in ad5791_get_lin_comp() 214 else if (span <= 12000) in ad5791_get_lin_comp() 216 else if (span <= 16000) in ad5791_get_lin_comp() 218 else if (span <= 19000) in ad5791_get_lin_comp() 224 static int ad5780_get_lin_comp(unsigned int span) in ad5780_get_lin_comp() argument 226 if (span <= 10000) in ad5780_get_lin_comp()
|
/Linux-v5.10/arch/sh/mm/ |
D | pmb.c | 146 unsigned long span; in pmb_mapping_exists() local 169 span = pmbe->size; in pmb_mapping_exists() 176 span += iter->size; in pmb_mapping_exists() 181 if (size <= span) { in pmb_mapping_exists() 381 * Link adjacent entries that span multiple PMB in pmb_bolt_mapping() 640 * see if the entries span a contiguous mapping. If so, in pmb_synchronize() 657 unsigned long span, newsize; in pmb_merge() local 661 span = newsize = head->size; in pmb_merge() 665 span += tail->size; in pmb_merge() 667 if (pmb_size_valid(span)) { in pmb_merge() [all …]
|
/Linux-v5.10/Documentation/sphinx/ |
D | rstFlatTable.py | 17 * *column-span*: with the role ``cspan`` a cell can be extended through 20 * *row-span*: with the role ``rspan`` a cell can be extended through 23 * *auto span* rightmost cell of a table row over the missing cells on the 25 can changed from *auto span* to *auto fill*, which automaticly inserts 87 nodelist = [colSpan(span=int(text))] 98 nodelist = [rowSpan(span=int(text))] 367 cspan = elem.get("span") 371 rspan = elem.get("span")
|
/Linux-v5.10/lib/ |
D | glob.c | 76 * Iterate over each span in the character class. in glob_match() 77 * A span is either a single character a, or a in glob_match() 78 * range a-b. The first span may begin with ']'. in glob_match()
|
/Linux-v5.10/tools/perf/ui/gtk/ |
D | browser.c | 39 return "<span fgcolor='red'>"; in perf_gtk__get_percent_color() 41 return "<span fgcolor='dark green'>"; in perf_gtk__get_percent_color()
|
D | annotate.c | 52 ret += scnprintf(buf + ret, size - ret, "</span>"); in perf_gtk__get_percent() 74 const char *markup = "<span fgcolor='gray'>"; in perf_gtk__get_line() 88 ret += scnprintf(buf + ret, size - ret, "</span>"); in perf_gtk__get_line()
|
/Linux-v5.10/net/mac802154/ |
D | rx.c | 37 __le16 span, sshort; in ieee802154_subif_frame() local 42 span = wpan_dev->pan_id; in ieee802154_subif_frame() 55 if (mac_cb(skb)->dest.pan_id != span && in ieee802154_subif_frame() 64 if (mac_cb(skb)->dest.pan_id != span && in ieee802154_subif_frame()
|
/Linux-v5.10/mm/ |
D | shuffle.c | 72 * the span, end_pfn - start_pfn. 99 * page_j randomly selected in the span @zone_start_pfn to in __shuffle_zone() 108 * Pick a random order aligned page in the zone span as in __shuffle_zone()
|
/Linux-v5.10/include/linux/sched/ |
D | topology.h | 147 * Span of all CPUs in this domain. 153 unsigned long span[]; member 158 return to_cpumask(sd->span); in sched_domain_span()
|
/Linux-v5.10/kernel/sched/ |
D | topology.c | 44 printk(KERN_CONT "span=%*pbl level=%s\n", in sched_domain_debug_one() 48 printk(KERN_ERR "ERROR: domain->span does not contain CPU%d\n", cpu); in sched_domain_debug_one() 92 printk(KERN_CONT " %d:{ span=%*pbl", in sched_domain_debug_one() 122 printk(KERN_ERR "ERROR: groups don't span domain->span\n"); in sched_domain_debug_one() 126 printk(KERN_ERR "ERROR: parent span is not a superset of domain->span\n"); in sched_domain_debug_one() 447 free_cpumask_var(rd->span); in free_rootdomain() 465 cpumask_clear_cpu(rq->cpu, old_rd->span); in rq_attach_root() 479 cpumask_set_cpu(rq->cpu, rd->span); in rq_attach_root() 504 if (!zalloc_cpumask_var(&rd->span, GFP_KERNEL)) in init_rootdomain() 536 free_cpumask_var(rd->span); in init_rootdomain() [all …]
|
/Linux-v5.10/Documentation/sphinx-static/ |
D | theme_overrides.css | 106 span.menuselection { 111 code.kbd, code.kbd span {
|
/Linux-v5.10/drivers/perf/ |
D | arm_smmuv3_pmu.c | 251 int idx, u32 span, u32 sid) in smmu_pmu_set_event_filter() argument 256 evtyper = get_event(event) | span << SMMU_PMCG_SID_SPAN_SHIFT; in smmu_pmu_set_event_filter() 277 u32 span, sid; in smmu_pmu_apply_event_filter() local 281 span = filter_en ? get_filter_span(event) : in smmu_pmu_apply_event_filter() 288 smmu_pmu_set_event_filter(event, idx, span, sid); in smmu_pmu_apply_event_filter() 296 smmu_pmu_set_event_filter(event, 0, span, sid); in smmu_pmu_apply_event_filter()
|
/Linux-v5.10/tools/testing/selftests/drivers/net/mlxsw/ |
D | qos_headroom.sh | 330 # Egress SPAN should have added to the "invisible" buffer configuration. 348 # After SPAN removal, hidden buffers should be back to the original sizes. 353 check_err $? "SPAN removed: Invisible buffers account for '$dsize_4', expected '== $dsize'"
|
/Linux-v5.10/Documentation/doc-guide/ |
D | sphinx.rst | 267 * column-span: with the role ``cspan`` a cell can be extended through 270 * row-span: with the role ``rspan`` a cell can be extended through 273 * auto span rightmost cell of a table row over the missing cells on the right 275 changed from *auto span* to *auto fill*, which automatically inserts (empty)
|
/Linux-v5.10/Documentation/translations/it_IT/doc-guide/ |
D | sphinx.rst | 293 * column-span: col ruolo ``cspan`` una cella può essere estesa attraverso 296 * raw-span: col ruolo ``rspan`` una cella può essere estesa attraverso 299 * auto-span: la cella più a destra viene estesa verso destra per compensare 301 può essere cambiato da *auto-span* ad *auto-fill*, il quale inserisce
|