Home
last modified time | relevance | path

Searched +full:spin +full:- +full:up (Results 1 – 25 of 369) sorted by relevance

12345678910>>...15

/Linux-v6.1/Documentation/hwmon/
Dlm93.rst10 Addresses scanned: I2C 0x2c-0x2e
18 Addresses scanned: I2C 0x2c-0x2e
24 - Mark M. Hoffman <mhoffman@lightlink.com>
25 - Ported to 2.6 by Eric J. Bowersox <ericb@aspsys.com>
26 - Adapted to 2.6.20 by Carsten Emde <ce@osadl.org>
27 - Modified for mainline integration by Hans J. Koch <hjk@hansjkoch.de>
30 -----------------
33 Set to non-zero to force some initializations (default is 0).
38 Configures in7 and in8 limit type, where 0 means absolute and non-zero
54 --------------------
[all …]
Ddrivetemp.rst1 .. SPDX-License-Identifier: GPL-2.0
8 ----------
10 ANS T13/1699-D
11 Information technology - AT Attachment 8 - ATA/ATAPI Command Set (ATA8-ACS)
14 Information technology - SCSI Primary Commands - 4 (SPC-4)
17 Information technology - SCSI / ATA Translation - 5 (SAT-5)
21 -----------
34 ----------
36 Reading the drive temperature may reset the spin down timer on some drives.
43 change its mode (meaning the drive will not spin up). It is unknown if other
[all …]
/Linux-v6.1/tools/testing/selftests/kvm/
Dset_memory_region_test.c1 // SPDX-License-Identifier: GPL-2.0
56 struct kvm_run *run = vcpu->run; in vcpu_worker()
61 * Loop until the guest is done. Re-enter the guest on all MMIO exits, in vcpu_worker()
68 if (run->exit_reason == KVM_EXIT_IO) { in vcpu_worker()
77 if (run->exit_reason != KVM_EXIT_MMIO) in vcpu_worker()
80 TEST_ASSERT(!run->mmio.is_write, "Unexpected exit mmio write"); in vcpu_worker()
81 TEST_ASSERT(run->mmio.len == 8, in vcpu_worker()
82 "Unexpected exit mmio size = %u", run->mmio.len); in vcpu_worker()
84 TEST_ASSERT(run->mmio.phys_addr == MEM_REGION_GPA, in vcpu_worker()
86 run->mmio.phys_addr); in vcpu_worker()
[all …]
/Linux-v6.1/drivers/gpu/drm/i915/gt/
Dselftest_rps.c1 // SPDX-License-Identifier: MIT
25 #define CPU_LATENCY 0 /* -1 to disable pm_qos, 0 to disable cstates */
36 return -1; in cmp_u64()
48 return -1; in cmp_u32()
67 #define CS_GPR(x) GEN8_RING_CS_GPR(engine->mmio_base, x) in create_spin_counter()
75 obj = i915_gem_object_create_internal(vm->i915, 64 << 10); in create_spin_counter()
79 end = obj->base.size / sizeof(u32) - 1; in create_spin_counter()
112 loop = cs - base; in create_spin_counter()
125 *cs++ = lower_32_bits(vma->node.start + end * sizeof(*cs)); in create_spin_counter()
126 *cs++ = upper_32_bits(vma->node.start + end * sizeof(*cs)); in create_spin_counter()
[all …]
Dselftest_engine_pm.c1 // SPDX-License-Identifier: GPL-2.0
24 return *a - *b; in cmp_u64()
75 struct intel_engine_cs *engine = ce->engine; in __measure_timestamps()
76 u32 *sema = memset32(engine->status_page.addr + 1000, 0, 5); in __measure_timestamps()
77 u32 offset = i915_ggtt_offset(engine->status_page.vma); in __measure_timestamps()
95 cs = emit_srm(cs, RING_TIMESTAMP(engine->mmio_base), offset + 4000); in __measure_timestamps()
96 cs = emit_srm(cs, RING_CTX_TIMESTAMP(engine->mmio_base), offset + 4004); in __measure_timestamps()
101 cs = emit_srm(cs, RING_TIMESTAMP(engine->mmio_base), offset + 4016); in __measure_timestamps()
102 cs = emit_srm(cs, RING_CTX_TIMESTAMP(engine->mmio_base), offset + 4012); in __measure_timestamps()
116 while (READ_ONCE(sema[1]) == 0) /* wait for the gpu to catch up */ in __measure_timestamps()
[all …]
Dselftest_execlists.c1 // SPDX-License-Identifier: MIT
24 #define CS_GPR(engine, n) ((engine)->mmio_base + 0x600 + (n) * 4)
47 tasklet_hi_schedule(&engine->sched_engine->tasklet); in wait_for_submit()
58 if (!READ_ONCE(engine->execlists.pending[0]) && is_active(rq)) in wait_for_submit()
62 return -ETIME; in wait_for_submit()
78 if (READ_ONCE(engine->execlists.pending[0])) in wait_for_reset()
84 if (READ_ONCE(rq->fence.error)) in wait_for_reset()
90 if (rq->fence.error != -EIO) { in wait_for_reset()
92 engine->name, in wait_for_reset()
93 rq->fence.context, in wait_for_reset()
[all …]
/Linux-v6.1/drivers/of/unittest-data/
Doverlay.dts1 // SPDX-License-Identifier: GPL-2.0
2 /dts-v1/;
9 hvac_2: hvac-large-1 {
10 compatible = "ot,hvac-large";
11 heat-range = < 40 75 >;
12 cool-range = < 65 80 >;
18 #address-cells = <1>;
19 #size-cells = <1>;
23 #address-cells = <1>;
24 #size-cells = <1>;
[all …]
/Linux-v6.1/arch/sparc/include/asm/
Dbackoff.h1 /* SPDX-License-Identifier: GPL-2.0 */
10 * completion of the compare-and-swap instruction. Heavily
14 * When an atomic operation fails and needs to be retried, we spin a
16 * operation we double the spin count, realizing an exponential
19 * When we spin, we try to use an operation that will cause the
24 * On all cpus prior to SPARC-T4 we do three dummy reads of the
28 * For SPARC-T4 and later we have a special "pause" instruction
31 * unless a disrupting trap happens first. SPARC-T4 specifically
39 * on earlier chips, we shift the backoff value up by 7 bits. (Three
/Linux-v6.1/drivers/net/can/softing/
Dsofting_main.c1 // SPDX-License-Identifier: GPL-2.0-only
3 * Copyright (C) 2008-2010
5 * - Kurt Van Dijck, EIA Electronics
15 #define TX_ECHO_SKB_MAX (((TXMAX+1)/2)-1)
19 * is online (ie. up 'n running, not sleeping, not busoff
27 return (can->state <= CAN_STATE_ERROR_PASSIVE); in canif_is_active()
33 if (card->pdat->generation >= 2) { in softing_set_reset_dpram()
34 spin_lock_bh(&card->spin); in softing_set_reset_dpram()
35 iowrite8(ioread8(&card->dpram[DPRAM_V2_RESET]) & ~1, in softing_set_reset_dpram()
36 &card->dpram[DPRAM_V2_RESET]); in softing_set_reset_dpram()
[all …]
/Linux-v6.1/kernel/locking/
Dspinlock.c1 // SPDX-License-Identifier: GPL-2.0
10 * SMP and the DEBUG_SPINLOCK cases. (UP-nondebug inlines them)
33 * If lockdep is enabled then we use the non-preemption spin-ops
35 * not re-enabled during lock-acquire (which the preempt-spin-ops do):
63 * This could be a long-held lock. We both prepare to spin for a long
76 arch_##op##_relax(&lock->raw_lock); \
92 arch_##op##_relax(&lock->raw_lock); \
109 /* irq-disabling. We use the generic preemption-aware */ \
118 * Build preemption-friendly versions of the following
119 * lock-spinning functions:
[all …]
Dspinlock_rt.c1 // SPDX-License-Identifier: GPL-2.0-only
3 * PREEMPT_RT substitution for spin/rw_locks
8 * - Contrary to plain rtmutexes, spinlocks and rwlocks are state
11 * during that time are redirected to the saved state so no wake up is
14 * - Non RT spin/rwlocks disable preemption and eventually interrupts.
29 * preserving. Take RCU nesting into account as spin/read/write_lock() can
47 rtlock_lock(&lock->lock); in __rt_spin_lock()
54 spin_acquire(&lock->dep_map, 0, 0, _RET_IP_); in rt_spin_lock()
62 spin_acquire(&lock->dep_map, subclass, 0, _RET_IP_); in rt_spin_lock_nested()
70 spin_acquire_nest(&lock->dep_map, 0, 0, nest_lock, _RET_IP_); in rt_spin_lock_nest_lock()
[all …]
/Linux-v6.1/arch/x86/include/asm/
Dspinlock.h1 /* SPDX-License-Identifier: GPL-2.0 */
16 * Simple spin lock operations. There are two variants, one clears IRQ's
19 * These are fair FIFO ticket locks, which support up to 2^16 CPUs.
24 /* How long a lock should spin before we consider blocking */
30 * Read-write spinlocks, allowing multiple readers
35 * can "mix" irq-safe locks - any writer needs to get a
36 * irq-safe write-lock, but readers can get non-irqsafe
37 * read-locks.
39 * On x86, we implement read-write locks using the generic qrwlock with
/Linux-v6.1/Documentation/locking/
Dspinlocks.rst5 Lesson 1: Spin locks
20 there is only one thread-of-control within the region(s) protected by that
21 lock. This works well even under UP also, so the code does _not_ need to
22 worry about UP vs SMP issues: the spinlocks work correctly under both.
26 Documentation/memory-barriers.txt
33 spinlock for most things - using more than one spinlock can make things a
35 sequences that you **know** need to be split up: avoid it at all cost if you
45 NOTE! The spin-lock is safe only when you **also** use the lock itself
50 ----
52 Lesson 2: reader-writer spinlocks.
[all …]
Drt-mutex-design.rst2 RT-mutex implementation design
12 Documentation/locking/rt-mutex.rst. Although this document does explain problems
22 ----------------------------
49 A ---+
52 C +----+
54 B +-------->
59 -------------------------
74 -----------
80 - The PI chain is an ordered series of locks and processes that cause
86 - In this document, to differentiate from locks that implement
[all …]
Dmutex-design.rst10 -----------------
24 --------------
28 (->owner) to keep track of the lock state during its lifetime. Field owner
32 if waiter list is non-empty). In its most basic form it also includes a
33 wait-queue and a spinlock that serializes access to it. Furthermore,
34 CONFIG_MUTEX_SPIN_ON_OWNER=y systems use a spinner MCS lock (->osq), described
45 (ii) midpath: aka optimistic spinning, tries to spin for acquisition
49 soon. The mutex spinners are queued up using MCS lock so that only
52 The MCS lock (proposed by Mellor-Crummey and Scott) is a simple spinlock
55 cacheline bouncing that common test-and-set spinlock implementations
[all …]
/Linux-v6.1/arch/arc/include/asm/
Datomic-spinlock.h1 /* SPDX-License-Identifier: GPL-2.0-only */
7 * Non hardware assisted Atomic-R-M-W
8 * Locking would change to irq-disabling only (UP) and spinlocks (SMP)
25 WRITE_ONCE(v->counter, i); in arch_atomic_set()
37 v->counter c_op i; \
48 * spin lock/unlock provides the needed smp_mb() before/after \
51 temp = v->counter; \
53 v->counter = temp; \
66 * spin lock/unlock provides the needed smp_mb() before/after \
69 orig = v->counter; \
[all …]
/Linux-v6.1/arch/arm64/kernel/
Dsmp_spin_table.c1 // SPDX-License-Identifier: GPL-2.0-only
3 * Spin Table SMP initialisation
50 return -ENODEV; in smp_spin_table_cpu_init()
55 ret = of_property_read_u64(dn, "cpu-release-addr", in smp_spin_table_cpu_init()
58 pr_err("CPU %d: missing or invalid cpu-release-addr property\n", in smp_spin_table_cpu_init()
72 return -ENODEV; in smp_spin_table_cpu_prepare()
75 * The cpu-release-addr may or may not be inside the linear mapping. in smp_spin_table_cpu_prepare()
83 return -ENOMEM; in smp_spin_table_cpu_prepare()
87 * endianness of the kernel. Therefore, any boot-loaders that in smp_spin_table_cpu_prepare()
89 * boot-loader's endianness before jumping. This is mandated by in smp_spin_table_cpu_prepare()
[all …]
/Linux-v6.1/drivers/gpu/drm/i915/gem/selftests/
Di915_gem_context.c2 * SPDX-License-Identifier: MIT
42 int err = -ENODEV; in live_nop_switch()
52 if (!DRIVER_CAPS(i915)->has_logical_contexts) in live_nop_switch()
61 err = -ENOMEM; in live_nop_switch()
88 i915_request_await_dma_fence(this, &rq->fence); in live_nop_switch()
98 err = -EIO; in live_nop_switch()
106 nctx, engine->name, ktime_to_ns(times[1] - times[0])); in live_nop_switch()
108 err = igt_live_test_begin(&t, i915, __func__, engine->name); in live_nop_switch()
127 i915_request_await_dma_fence(this, &rq->fence); in live_nop_switch()
137 * and submission of our breadcrumbs - in live_nop_switch()
[all …]
/Linux-v6.1/tools/testing/selftests/rcutorture/bin/
Djitterstart.sh2 # SPDX-License-Identifier: GPL-2.0+
4 # Start up the specified number of jitter.sh scripts in the background.
6 # Usage: . jitterstart.sh n jittering-dir duration [ sleepmax [ spinmax ] ]
8 # n: Number of jitter.sh scripts to start up.
9 # jittering-dir: Directory in which to put "jittering" file.
12 # spinmax: Maximum microseconds to spin, defaults to one millisecond.
19 if test -z "$jitter_n"
25 if test -z "$jittering_dir"
/Linux-v6.1/Documentation/admin-guide/laptops/
Dlaptop-mode.rst2 How to conserve battery power using laptop-mode
12 ------------
14 Laptop mode is used to minimize the time that the hard disk needs to be spun up,
31 ------------
41 located in /etc/default/laptop-mode on Debian-based systems, or in
42 /etc/sysconfig/laptop-mode on other systems.
52 -------
54 * The downside of laptop mode is that you have a chance of losing up to 10
64 * If you mount some of your ext3/reiserfs filesystems with the -n option, then
67 wrong options -- or it will fail because it cannot write to /etc/mtab.
[all …]
/Linux-v6.1/Documentation/trace/
Dhwlat_detector.rst6 -------------
17 even know that they are occuring. SMIs are instead set up by BIOS code
36 ------
50 - width - time period to sample with CPUs held (usecs)
52 - window - total period of sampling, width being inside (usecs)
55 for every 1,000,000 usecs (1s) the hwlat detector will spin for 500,000 usecs
74 - tracing_threshold - minimum latency value to be considered (usecs)
75 - tracing_max_latency - maximum hardware latency actually observed (usecs)
76 - tracing_cpumask - the CPUs to move the hwlat thread across
77 - hwlat_detector/width - specified amount of time to spin within window (usecs)
[all …]
/Linux-v6.1/Documentation/driver-api/thermal/
Dintel_powerclamp.rst6 - Arjan van de Ven <arjan@linux.intel.com>
7 - Jacob Pan <jacob.jun.pan@linux.intel.com>
12 - Goals and Objectives
15 - Idle Injection
16 - Calibration
19 - Effectiveness and Limitations
20 - Power vs Performance
21 - Scalability
22 - Calibration
23 - Comparison with Alternative Techniques
[all …]
/Linux-v6.1/arch/powerpc/include/asm/
Dparavirt.h1 /* SPDX-License-Identifier: GPL-2.0-or-later */
49 * yielded). So this may not be a problem for simple spin locks because the
52 * However the queued spin lock contended path is more strictly ordered: the
54 * recursing on that lock will cause the task to then queue up again behind the
71 plpar_hcall_norets_notrace(H_CONFER, -1, 0); in yield_to_any()
132 * we're not accessing per-cpu resources in a way that can in vcpu_is_preempted()
/Linux-v6.1/arch/arm/mach-omap2/
Dsram242x.S1 /* SPDX-License-Identifier: GPL-2.0-or-later */
3 * linux/arch/arm/mach-omap2/sram242x.S
9 * Richard Woodruff <r-woodruff2@ti.com>
31 stmfd sp!, {r0 - r12, lr} @ save registers on stack
39 str r3, [r2] @ go to L1-freq operation
42 mov r9, #0x1 @ set up for L1 voltage call
61 /* voltage shift up */
62 mov r9, #0x0 @ shift back to L0-voltage
65 /* frequency shift up */
67 str r3, [r2] @ go to L0-freq operation
[all …]
Dsram243x.S1 /* SPDX-License-Identifier: GPL-2.0-or-later */
3 * linux/arch/arm/mach-omap2/sram243x.S
9 * Richard Woodruff <r-woodruff2@ti.com>
31 stmfd sp!, {r0 - r12, lr} @ save registers on stack
39 str r3, [r2] @ go to L1-freq operation
42 mov r9, #0x1 @ set up for L1 voltage call
61 /* voltage shift up */
62 mov r9, #0x0 @ shift back to L0-voltage
65 /* frequency shift up */
67 str r3, [r2] @ go to L0-freq operation
[all …]

12345678910>>...15