/Linux-v5.4/kernel/sched/ |
D | loadavg.c | 72 void get_avenrun(unsigned long *loads, unsigned long offset, int shift) in get_avenrun() argument 74 loads[0] = (avenrun[0] + offset) << shift; in get_avenrun() 75 loads[1] = (avenrun[1] + offset) << shift; in get_avenrun() 76 loads[2] = (avenrun[2] + offset) << shift; in get_avenrun()
|
/Linux-v5.4/arch/powerpc/perf/ |
D | power9-pmu.c | 160 GENERIC_EVENT_ATTR(mem-loads, MEM_LOADS); 164 CACHE_EVENT_ATTR(L1-dcache-loads, PM_LD_REF_L1); 168 CACHE_EVENT_ATTR(L1-icache-loads, PM_INST_FROM_L1); 171 CACHE_EVENT_ATTR(LLC-loads, PM_DATA_FROM_L3); 174 CACHE_EVENT_ATTR(branch-loads, PM_BR_CMPL);
|
D | power8-pmu.c | 134 CACHE_EVENT_ATTR(L1-dcache-loads, PM_LD_REF_L1); 139 CACHE_EVENT_ATTR(L1-icache-loads, PM_INST_FROM_L1); 143 CACHE_EVENT_ATTR(LLC-loads, PM_DATA_FROM_L3); 149 CACHE_EVENT_ATTR(branch-loads, PM_BRU_FIN);
|
/Linux-v5.4/tools/perf/Documentation/ |
D | perf-mem.txt | 19 right set of options to display a memory access profile. By default, loads 20 and stores are sampled. Use the -t option to limit to loads or stores. 85 Specify desired latency for loads event. (x86 only)
|
D | perf-c2c.txt | 52 Configure mem-loads latency. (x86 only) 127 cpu/mem-loads,ldlat=30/P 132 cpu/mem-loads/
|
/Linux-v5.4/arch/alpha/lib/ |
D | ev6-copy_user.S | 64 EXI( ldbu $1,0($17) ) # .. .. .. L : Keep loads separate from stores 116 EXI ( ldbu $2,0($17) ) # .. .. .. L : No loads in the same quad 203 EXI ( ldbu $2,0($17) ) # .. .. .. L : No loads in the same quad
|
/Linux-v5.4/include/uapi/linux/ |
D | sysinfo.h | 10 __kernel_ulong_t loads[3]; /* 1, 5, and 15 minute load averages */ member
|
/Linux-v5.4/include/linux/sched/ |
D | loadavg.h | 16 extern void get_avenrun(unsigned long *loads, unsigned long offset, int shift);
|
/Linux-v5.4/Documentation/x86/ |
D | tsx_async_abort.rst | 13 case certain loads may speculatively pass invalid data to dependent operations 15 Synchronization Extensions (TSX) transaction. This includes loads with no 16 fault or assist condition. Such loads may speculatively expose stale data from
|
/Linux-v5.4/Documentation/devicetree/bindings/remoteproc/ |
D | stm32-rproc.txt | 3 This document defines the binding for the remoteproc component that loads and 51 - st,auto-boot: If defined, when remoteproc is probed, it loads the default
|
/Linux-v5.4/tools/memory-model/Documentation/ |
D | explanation.txt | 78 for the loads, the model will predict whether it is possible for the 79 code to run in such a way that the loads will indeed obtain the 141 shared memory locations and another CPU loads from those locations in 153 A memory model will predict what values P1 might obtain for its loads 196 Since r1 = 1, P0 must store 1 to flag before P1 loads 1 from 197 it, as loads can obtain values only from earlier stores. 199 P1 loads from flag before loading from buf, since CPUs execute 222 each CPU stores to its own shared location and then loads from the 271 X: P1 loads 1 from flag executes before 272 Y: P1 loads 0 from buf executes before [all …]
|
D | recipes.txt | 46 tearing, load/store fusing, and invented loads and stores. 204 and another CPU execute a pair of loads from this same pair of variables, 311 smp_rmb() macro orders prior loads against later loads. Therefore, if 354 second, while another CPU loads from the second variable and then stores 475 that one CPU first stores to one variable and then loads from a second, 476 while another CPU stores to the second variable and then loads from the
|
/Linux-v5.4/Documentation/core-api/ |
D | refcount-vs-atomic.rst | 41 A strong (full) memory ordering guarantees that all prior loads and 49 A RELEASE memory ordering guarantees that all prior loads and 57 An ACQUIRE memory ordering guarantees that all post loads and
|
/Linux-v5.4/Documentation/ |
D | memory-barriers.txt | 180 perceived by the loads made by another CPU in the same order as the stores were 249 (*) Overlapping loads and stores within a particular CPU will appear to be 277 (*) It _must_not_ be assumed that independent loads and stores will be issued 371 deferral and combination of memory operations; speculative loads; speculative 390 to have any effect on loads. 403 where two loads are performed such that the second depends on the result 409 A data dependency barrier is a partial ordering on interdependent loads 410 only; it is not required to have any effect on stores, independent loads 411 or overlapping loads. 419 touched by the load will be perceptible to any loads issued after the data [all …]
|
/Linux-v5.4/arch/mips/include/asm/ |
D | mips-r2-to-r6-emul.h | 22 u64 loads; member
|
D | fpu_emulator.h | 26 unsigned long loads; member
|
/Linux-v5.4/Documentation/livepatch/ |
D | module-elf-format.rst | 94 relocation section(s) to the driver once it loads. 204 a patch module loads. For example, this is the case when livepatch patches 206 symbols are resolved simply when the target module loads. In any case, for 304 additional arch-specific tasks when a target module loads, such as applying 306 .altinstructions and .parainstructions sections when a target module loads. 318 module->klp_info, which is a klp_modinfo struct. When a livepatch module loads,
|
/Linux-v5.4/arch/mips/kernel/ |
D | mips-r2-to-r6-emul.c | 1274 MIPS_R2_STATS(loads); in mipsr2_decoder() 1348 MIPS_R2_STATS(loads); in mipsr2_decoder() 1608 MIPS_R2_STATS(loads); in mipsr2_decoder() 1727 MIPS_R2_STATS(loads); in mipsr2_decoder() 2267 (unsigned long)__this_cpu_read(mipsr2emustats.loads), in mipsr2_emul_show() 2268 (unsigned long)__this_cpu_read(mipsr2bdemustats.loads)); in mipsr2_emul_show() 2324 __this_cpu_write((mipsr2emustats).loads, 0); in mipsr2_clear_show() 2325 __this_cpu_write((mipsr2bdemustats).loads, 0); in mipsr2_clear_show()
|
/Linux-v5.4/kernel/debug/kdb/ |
D | kdb_main.c | 2498 val->loads[0] = avenrun[0]; in kdb_sysinfo() 2499 val->loads[1] = avenrun[1]; in kdb_sysinfo() 2500 val->loads[2] = avenrun[2]; in kdb_sysinfo() 2544 LOAD_INT(val.loads[0]), LOAD_FRAC(val.loads[0]), in kdb_summary() 2545 LOAD_INT(val.loads[1]), LOAD_FRAC(val.loads[1]), in kdb_summary() 2546 LOAD_INT(val.loads[2]), LOAD_FRAC(val.loads[2])); in kdb_summary()
|
/Linux-v5.4/tools/testing/selftests/livepatch/ |
D | README | 7 The test suite loads and unloads several test kernel modules to verify
|
/Linux-v5.4/tools/testing/selftests/powerpc/copyloops/ |
D | memcpy_64.S | 115 ld r9,0(r4) # 3+2n loads, 2+2n stores 127 0: ld r0,0(r4) # 4+2n loads, 3+2n stores
|
/Linux-v5.4/arch/powerpc/lib/ |
D | memcpy_64.S | 115 ld r9,0(r4) # 3+2n loads, 2+2n stores 127 0: ld r0,0(r4) # 4+2n loads, 3+2n stores
|
/Linux-v5.4/Documentation/ABI/stable/ |
D | vdso | 1 On some architectures, when the kernel loads any userspace program it
|
/Linux-v5.4/Documentation/admin-guide/LSM/ |
D | LoadPin.rst | 29 still use LoadPin to protect the integrity of other files kernel loads. The
|
/Linux-v5.4/tools/perf/util/ |
D | parse-events.l | 346 mem-loads { return str(yyscanner, PE_KERNEL_PMU_EVENT); } 358 load|loads|read |
|