/Linux-v4.19/drivers/net/ethernet/cavium/liquidio/ |
D | octeon_main.h | 153 wait_queue_entry_t we; in sleep_cond() local 155 init_waitqueue_entry(&we, current); in sleep_cond() 156 add_wait_queue(wait_queue, &we); in sleep_cond() 167 remove_wait_queue(wait_queue, &we); in sleep_cond() 180 wait_queue_entry_t we; in sleep_timeout_cond() local 182 init_waitqueue_entry(&we, current); in sleep_timeout_cond() 183 add_wait_queue(wait_queue, &we); in sleep_timeout_cond() 188 remove_wait_queue(wait_queue, &we); in sleep_timeout_cond()
|
/Linux-v4.19/Documentation/devicetree/bindings/pinctrl/ |
D | sprd,pinctrl.txt | 12 to choose one function (like: UART0) for which system, since we 15 There are too much various configuration that we can not list all 16 of them, so we can not make every Spreadtrum-special configuration 18 global configuration in future. Then we add one "sprd,control" to 19 set these various global control configuration, and we need use 22 Moreover we recognise every fields comprising one bit or several 23 bits in one global control register as one pin, thus we should 32 Now we have 4 systems for sleep mode on SC9860 SoC: AP system, 42 In some situation we need set the pin sleep mode and pin sleep related 45 sleep mode. For example, if we set the pin sleep mode as PUBCP_SLEEP [all …]
|
/Linux-v4.19/Documentation/x86/ |
D | entry_64.txt | 52 so. If we mess that up even slightly, we crash. 54 So when we have a secondary entry, already in kernel mode, we *must 55 not* use SWAPGS blindly - nor must we forget doing a SWAPGS when it's 81 If we are at an interrupt or user-trap/gate-alike boundary then we can 83 whether SWAPGS was already done: if we see that we are a secondary 84 entry interrupting kernel mode execution, then we know that the GS 85 base has already been switched. If it says that we interrupted 86 user-space execution then we must do the SWAPGS. 88 But if we are in an NMI/MCE/DEBUG/whatever super-atomic entry context, 90 stack but before we executed SWAPGS, then the only safe way to check [all …]
|
/Linux-v4.19/fs/jffs2/ |
D | TODO | 5 reservations will necessarily be pessimistic. With this we could even 6 do shared writable mmap, if we can have a fs hook for do_wp_page() to 11 - checkpointing (do we need this? scan is quite fast) 25 nodes to a different one, we can separate clean nodes from those which 28 - Stop keeping name in-core with struct jffs2_full_dirent. If we keep the hash in 29 the full dirent, we only need to go to the flash in lookup() when we think we've
|
/Linux-v4.19/Documentation/filesystems/ |
D | xfs-delayed-logging-design.txt | 25 That is, if we have a sequence of changes A through to F, and the object was 26 written to disk after change D, we would see in the log the following series 91 relogging technique XFS uses is that we can be relogging changed objects 92 multiple times before they are committed to disk in the log buffers. If we 98 contains all the changes from the previous changes. In other words, we have one 100 wasting space. When we are doing repeated operations on the same set of 103 log would greatly reduce the amount of metadata we write to the log, and this 110 formatting the changes in a transaction to the log buffer. Hence we cannot avoid 113 Delayed logging is the name we've given to keeping and tracking transactional 163 changes to the log buffers, we need to ensure that the object we are formatting [all …]
|
D | directory-locking | 5 When taking the i_rwsem on multiple non-directory objects, we 11 1) read access. Locking rules: caller locks directory we are accessing. 24 lock it. If we need to lock both, lock them in inode pointer order. 26 NB: we might get away with locking the the source (and target in exchange 47 lock it. If we need to lock both, do so in inode pointer order. 49 All ->i_rwsem are taken exclusive. Again, we might get away with locking 59 First of all, at any moment we have a partial ordering of the 65 attempts to acquire lock on B, A will remain the parent of B until we 71 renames will be blocked on filesystem lock and we don't start changing 72 the order until we had acquired all locks). [all …]
|
D | path-lookup.txt | 49 the path given by the name's starting point (which we know in advance -- eg. 55 A parent, of course, must be a directory, and we must have appropriate 79 In order to lookup a dcache (parent, name) tuple, we take a hash on the tuple 81 in that bucket is then walked, and we do a full comparison of each entry 148 However, when inserting object 2 onto a new list, we end up with this: 161 Because we didn't wait for a grace period, there may be a concurrent lookup 182 As explained above, we would like to do path walking without taking locks or 188 than reloading from the dentry later on (otherwise we'd have interesting things 192 no non-atomic stores to shared data), and to recheck the seqcount when we are 194 Avoiding destructive or changing operations means we can easily unwind from [all …]
|
D | xfs-self-describing-metadata.txt | 28 However, if we scale the filesystem up to 1PB, we now have 10x as much metadata 40 magic number in the metadata block, we have no other way of identifying what it 53 Hence we need to record more information into the metadata to allow us to 55 of analysis. We can't protect against every possible type of error, but we can 62 hence parse and verify the metadata object. IF we can't independently identify 68 magic numbers. Hence we can change the on-disk format of all these objects to 72 self identifying and we can do much more expansive automated verification of the 76 integrity checking. We cannot trust the metadata if we cannot verify that it has 77 not been changed as a result of external influences. Hence we need some form of 79 block. If we can verify the block contains the metadata it was intended to [all …]
|
/Linux-v4.19/drivers/block/paride/ |
D | Transition-notes | 9 ps_spinlock. C is always preceded by B, since we can't reach it 10 other than through B and we don't drop ps_spinlock between them. 14 A and each B is preceded by either A or C. Moments when we enter 37 * in ps_tq_int(): from the moment when we get ps_spinlock() to the 73 we would have to be called for the PIA that got ->claimed_cont 83 it is holding pd_lock. The only place within the area where we 87 we were acquiring the lock, (1) would be already false, since 89 If it was 0 before we tried to acquire pd_lock, (2) would be 96 (4) is done the same way - all places where we release pi_spinlock within 100 in the area, under pi_spinlock and we do not release it until after leaving [all …]
|
/Linux-v4.19/arch/x86/realmode/rm/ |
D | trampoline_64.S | 53 # write marker for master knows we're running 100 * the MSR for this AP. If SME is active and we've gotten this far 102 * don't we'll eventually crash trying to execute encrypted 139 * EFER.LMA = 1). Now we want to jump in 64bit mode, to do that we use
|
/Linux-v4.19/drivers/scsi/aic7xxx/ |
D | aic79xx.seq | 85 * If we have completions stalled waiting for the qfreeze 109 * ENSELO is cleared by a SELDO, so we must test for SELDO 169 * Since this status did not consume a FIFO, we have to 170 * be a bit more dilligent in how we check for FIFOs pertaining 178 * count in the SCB. In this case, we allow the routine servicing 183 * we detect case 1, we will properly defer the post of the SCB 222 * bad SCSI status (currently only for underruns), we 223 * queue the SCB for normal completion. Otherwise, we 258 * If we have relatively few commands outstanding, don't 303 * one byte of lun information we support. [all …]
|
D | aic7xxx.seq | 52 * After starting the selection hardware, we check for reconnecting targets 54 * bus arbitration. The problem with this is that we must keep track of the 55 * SCB that we've already pulled from the QINFIFO and started the selection 56 * on just in case the reselection wins so that we can retry the selection at 104 * We have at least one queued SCB now and we don't have any 124 * before we completed the DMA operation. If it was, 211 /* The Target ID we were selected at */ 239 * Watch ATN closely now as we pull in messages from the 285 * we've got a failed selection and must transition to bus 333 * Reselection has been initiated by a target. Make a note that we've been [all …]
|
/Linux-v4.19/Documentation/virtual/kvm/ |
D | locking.txt | 33 tracking i.e. the SPTE_SPECIAL_MASK is set. That means we need to 37 caused by write-protect. That means we just need to change the W bit of the 40 What we use to avoid all the race is the SPTE_HOST_WRITEABLE bit and 47 On fast page fault path, we will use cmpxchg to atomically set the spte W 52 But we need carefully check these cases: 54 The mapping from gfn to pfn may be changed since we can only ensure the pfn 83 For direct sp, we can easily avoid it since the spte of direct sp is fixed 84 to gfn. For indirect sp, before we do cmpxchg, we call gfn_to_pfn_atomic() 91 Then, we can ensure the dirty bitmaps is correctly set for a gfn. 93 Currently, to simplify the whole things, we disable fast page fault for [all …]
|
/Linux-v4.19/Documentation/devicetree/bindings/usb/ |
D | generic.txt | 4 - maximum-speed: tells USB controllers we want to work up to a certain 9 - dr_mode: tells Dual-Role USB controllers that we want to work on a 14 - phy_type: tells USB controllers that we want to configure the core to support 26 - hnp-disable: tells OTG controllers we want to disable OTG HNP, normally HNP 29 - srp-disable: tells OTG controllers we want to disable OTG SRP, SRP is 31 - adp-disable: tells OTG controllers we want to disable OTG ADP, ADP is
|
/Linux-v4.19/Documentation/usb/ |
D | WUSB-Design-overview.txt | 192 So let's say we connect a dongle to the system: it is detected and 195 Now we have a real HWA device connected and 214 So assuming we have devices and we have agreed for a channel to connect 215 on (let's say 9), we put the new RC to beacon: 283 ID and tell the HC to use all that. Then we start it. This means the HC 309 the device. First we allocate a /fake port/ and assign an 310 unauthenticated address (128 to 255--what we really do is 314 So now we are in the reset path -- we know we have a non-yet enumerated 315 device with an unauthorized address; we ask user space to authenticate 316 (FIXME: not yet done, similar to bluetooth pairing), then we do the key [all …]
|
/Linux-v4.19/Documentation/RCU/ |
D | rculist_nulls.txt | 23 * reuse these object before the RCU grace period, we 26 if (obj->key != key) { // not the object we expected 87 * we need to make sure obj->key is updated before obj->next 98 Nothing special here, we can use a standard RCU hlist deletion. 112 With hlist_nulls we can avoid extra smp_rmb() in lockless_lookup() 115 For example, if we choose to store the slot number as the 'nulls' 116 end-of-list marker for each slot of the hash table, we can detect 120 is not the slot number, then we must restart the lookup at 135 if (obj->key != key) { // not the object we expected 142 * if the nulls value we got at the end of this lookup is [all …]
|
/Linux-v4.19/Documentation/block/ |
D | deadline-iosched.txt | 21 service time for a request. As we focus mainly on read latencies, this is 50 When we have to move requests from the io scheduler queue to the block 51 device dispatch queue, we always give a preference to reads. However, we 53 how many times we give preference to reads over writes. When that has been 54 done writes_starved number of times, we dispatch some writes based on the 69 that comes at basically 0 cost we leave that on. We simply disable the
|
/Linux-v4.19/Documentation/power/ |
D | freezing-of-tasks.txt | 17 we only consider hibernation, but the description also applies to suspend). 28 it loop until PF_FROZEN is cleared for it. Then, we say that the task is 75 - freezes all tasks (including kernel threads) because we can't freeze 79 - thaws only kernel threads; this is particularly useful if we need to do 81 userspace tasks, or if we want to postpone the thawing of userspace tasks 84 - thaws all tasks (including kernel threads) because we can't thaw userspace 95 IV. Why do we do that? 100 hibernation. At the moment we have no simple means of checkpointing 102 metadata on disks, we cannot bring them back to the state from before the 113 2. Next, to create the hibernation image we need to free a sufficient amount of [all …]
|
/Linux-v4.19/Documentation/powerpc/ |
D | pci_iov_resource_on_powernv.txt | 32 The following section provides a rough description of what we have on P8 43 For DMA, MSIs and inbound PCIe error messages, we have a table (in 48 - For DMA we then provide an entire address space for each PE that can 54 - For MSIs, we have two windows in the address space (one at the top of 82 reserved for MSIs but this is not a problem at this point; we just 84 ignores that however and will forward in that space if we try). 91 Now, this is the "main" window we use in Linux today (excluding 96 Ideally we would like to be able to have individual functions in PEs 107 bits which are not conveyed by PowerBus but we don't use this. 109 * Can be configured to be segmented. When not segmented, we can [all …]
|
/Linux-v4.19/Documentation/devicetree/bindings/i2c/ |
D | i2c-arb-gpio-challenge.txt | 24 - OUR_CLAIM: output from us signaling to other hosts that we want the bus 31 Let's say we want to claim the bus. We: 35 3. Check THEIR_CLAIMS. If none are asserted then the we have the bus and we are 44 - our-claim-gpio: The GPIO that we use to claim the bus. 51 - wait-retry-us: we'll attempt another claim after this many microseconds. 53 - wait-free-us: we'll give up after this many microseconds. Default is 50000 us.
|
/Linux-v4.19/Documentation/networking/ |
D | fib_trie.txt | 33 verify that they actually do match the key we are searching for. 68 fib_find_node(). Inserting a new node means we might have to run the 103 slower than the corresponding fib_hash function, as we have to walk the 120 trie, key segment by key segment, until we find a leaf. check_leaf() does 123 If we find a match, we are done. 125 If we don't find a match, we enter prefix matching mode. The prefix length, 127 and we backtrack upwards through the trie trying to find a longest matching 133 the child index until we find a match or the child index consists of nothing but 136 At this point we backtrack (t->stats.backtrack++) up the trie, continuing to 139 At this point we will repeatedly descend subtries to look for a match, and there
|
/Linux-v4.19/Documentation/x86/x86_64/ |
D | 5level-paging.txt | 35 To mitigate this, we are not going to allocate virtual address space 41 If hint address set above 47-bit, but MAP_FIXED is not specified, we try 43 occupied, we look for unmapped area in *full* address space, rather than 57 One important case we need to handle here is interaction with MPX. 58 MPX (without MAWA extension) cannot handle addresses above 47-bit, so we 59 need to make sure that MPX cannot be enabled we already have VMA above
|
/Linux-v4.19/drivers/staging/vc04_services/interface/vchi/ |
D | TODO | 4 some of the ones we want: 16 to manage these buffers as dmabufs so that we can zero-copy import 23 days. Once we have the set of VCHI-using drivers we want in tree, we
|
/Linux-v4.19/tools/perf/ |
D | builtin-timechart.c | 412 struct wake_event *we = zalloc(sizeof(*we)); in sched_wakeup() local 414 if (!we) in sched_wakeup() 417 we->time = timestamp; in sched_wakeup() 418 we->waker = waker; in sched_wakeup() 419 we->backtrace = backtrace; in sched_wakeup() 422 we->waker = -1; in sched_wakeup() 424 we->wakee = wakee; in sched_wakeup() 425 we->next = tchart->wake_events; in sched_wakeup() 426 tchart->wake_events = we; in sched_wakeup() 427 p = find_create_pid(tchart, we->wakee); in sched_wakeup() [all …]
|
/Linux-v4.19/arch/x86/ras/ |
D | Kconfig | 8 PFN overflows, we try to soft-offline that page as we take it to mean 10 be best if we don't use it anymore.
|