Home
last modified time | relevance | path

Searched full:queued (Results 1 – 25 of 1466) sorted by relevance

12345678910>>...59

/Linux-v6.1/include/asm-generic/
Dqrwlock.h40 * queued_read_trylock - try to acquire read lock of a queued rwlock
41 * @lock : Pointer to queued rwlock structure
59 * queued_write_trylock - try to acquire write lock of a queued rwlock
60 * @lock : Pointer to queued rwlock structure
75 * queued_read_lock - acquire read lock of a queued rwlock
76 * @lock: Pointer to queued rwlock structure
91 * queued_write_lock - acquire write lock of a queued rwlock
92 * @lock : Pointer to queued rwlock structure
105 * queued_read_unlock - release read lock of a queued rwlock
106 * @lock : Pointer to queued rwlock structure
[all …]
Dqspinlock.h3 * Queued spinlock
48 * @lock: Pointer to queued spinlock structure
63 * @lock: queued spinlock structure
78 * @lock : Pointer to queued spinlock structure
86 * queued_spin_trylock - try to acquire the queued spinlock
87 * @lock : Pointer to queued spinlock structure
104 * queued_spin_lock - acquire a queued spinlock
105 * @lock: Pointer to queued spinlock structure
120 * queued_spin_unlock - release a queued spinlock
121 * @lock : Pointer to queued spinlock structure
[all …]
/Linux-v6.1/security/integrity/ima/
Dima_queue_keys.c18 * right away or should be queued for processing later.
29 * If custom IMA policy is not loaded then keys queued up
40 * queued up in case custom IMA policy was not loaded.
49 * This function sets up a worker to free queued keys in case
107 bool queued = false; in ima_queue_key() local
117 queued = true; in ima_queue_key()
121 if (!queued) in ima_queue_key()
124 return queued; in ima_queue_key()
128 * ima_process_queued_keys() - process keys queued for measurement
130 * This function sets ima_process_keys to true and processes queued keys.
[all …]
/Linux-v6.1/block/
Dblk-throttle.h14 * To avoid such starvation, dispatched bios are queued separately
19 * throtl_qnode is used to keep the queued bios separated by their sources.
20 * Bios are queued to throtl_qnode which in turn is queued to
24 * belongs to a throtl_grp and gets queued on itself or the parent, so
26 * queued and decrementing when dequeued is enough to keep the whole blkg
30 struct list_head node; /* service_queue->queued[] */
31 struct bio_list bios; /* queued bios */
39 * Bios queued directly to this service_queue or dispatched from
42 struct list_head queued[2]; /* throtl_qnode [READ/WRITE] */ member
43 unsigned int nr_queued[2]; /* number of queued bios */
[all …]
/Linux-v6.1/drivers/gpu/drm/
Ddrm_flip_work.c63 list_add_tail(&task->node, &work->queued); in drm_flip_work_queue_task()
92 * drm_flip_work_commit - commit queued work
94 * @wq: the work-queue to run the queued work on
96 * Trigger work previously queued by drm_flip_work_queue() to run
99 * prior), and then from vblank irq commit the queued work.
107 list_splice_tail(&work->queued, &work->commited); in drm_flip_work_commit()
108 INIT_LIST_HEAD(&work->queued); in drm_flip_work_commit()
151 INIT_LIST_HEAD(&work->queued); in drm_flip_work_init()
168 WARN_ON(!list_empty(&work->queued) || !list_empty(&work->commited)); in drm_flip_work_cleanup()
/Linux-v6.1/drivers/net/wireless/mediatek/mt76/
Ddebugfs.c59 seq_puts(s, " queue | hw-queued | head | tail |\n"); in mt76_queues_read()
67 i, q->queued, q->head, q->tail); in mt76_queues_read()
77 int i, queued; in mt76_rx_queues_read() local
79 seq_puts(s, " queue | hw-queued | head | tail |\n"); in mt76_rx_queues_read()
83 queued = mt76_is_usb(dev) ? q->ndesc - q->queued : q->queued; in mt76_rx_queues_read()
85 i, queued, q->head, q->tail); in mt76_rx_queues_read()
/Linux-v6.1/Documentation/userspace-api/media/mediactl/
Dmedia-request-ioc-queue.rst34 If the request was successfully queued, then the file descriptor can be
37 If the request was already queued before, then ``EBUSY`` is returned.
42 Once a request is queued, then the driver is required to gracefully handle
49 queued directly and you next try to queue a request, or vice versa.
62 The request was already queued or the application queued the first
Drequest-api.rst24 buffer queues since in practice only one buffer would be queued at a time.
59 instead of being immediately applied, and buffers queued to a request do not
60 enter the regular buffer queue until the request itself is queued.
66 queued by calling :ref:`MEDIA_REQUEST_IOC_QUEUE` on the request file descriptor.
68 A queued request cannot be modified anymore.
86 a buffer was queued via a request or vice versa will result in an ``EBUSY``
109 request that has been queued but not yet completed will return ``EBUSY``
121 longer in use by the kernel. That is, if the request is queued and then the
165 Once the request is fully prepared, it can be queued to the driver:
245 Once the request is fully prepared, it can be queued to the driver:
/Linux-v6.1/include/linux/
Ddynamic_queue_limits.h11 * 1) Objects are queued up to some limit specified as number of objects.
14 * 3) Starvation occurs when limit has been reached, all queued data has
17 * 4) Minimizing the amount of queued data is desirable.
24 * dql_avail - returns how many objects are available to be queued based
45 unsigned int num_queued; /* Total ever queued */
72 * Record number of objects queued. Assumes that caller has already checked
91 /* Returns how many objects can be queued, < 0 indicates over limit. */
/Linux-v6.1/net/x25/
Dx25_in.c210 int queued = 0; in x25_state3_machine() local
277 queued = 1; in x25_state3_machine()
315 queued = !sock_queue_rcv_skb(sk, skb); in x25_state3_machine()
319 queued = 1; in x25_state3_machine()
330 return queued; in x25_state3_machine()
418 int queued = 0, frametype, ns, nr, q, d, m; in x25_process_rx_frame() local
427 queued = x25_state1_machine(sk, skb, frametype); in x25_process_rx_frame()
430 queued = x25_state2_machine(sk, skb, frametype); in x25_process_rx_frame()
433 queued = x25_state3_machine(sk, skb, frametype, ns, nr, q, d, m); in x25_process_rx_frame()
436 queued = x25_state4_machine(sk, skb, frametype); in x25_process_rx_frame()
[all …]
/Linux-v6.1/net/rose/
Drose_in.c105 int queued = 0; in rose_state3_machine() local
168 queued = 1; in rose_state3_machine()
205 return queued; in rose_state3_machine()
266 int queued = 0, frametype, ns, nr, q, d, m; in rose_process_rx_frame() local
275 queued = rose_state1_machine(sk, skb, frametype); in rose_process_rx_frame()
278 queued = rose_state2_machine(sk, skb, frametype); in rose_process_rx_frame()
281 queued = rose_state3_machine(sk, skb, frametype, ns, nr, q, d, m); in rose_process_rx_frame()
284 queued = rose_state4_machine(sk, skb, frametype); in rose_process_rx_frame()
287 queued = rose_state5_machine(sk, skb, frametype); in rose_process_rx_frame()
293 return queued; in rose_process_rx_frame()
/Linux-v6.1/Documentation/core-api/
Dworkqueue.rst25 When a new work item gets queued, the worker begins executing again.
81 off of the queue, one after the other. If no work is queued, the
87 which manages worker-pools and processes the queued work items.
91 worker-pools to serve work items queued on unbound workqueues - the
102 When a work item is queued to a workqueue, the target worker-pool is
106 be queued on the worklist of either normal or highpri worker-pool that
143 on code paths that handle memory reclaim are required to be queued on
168 Work items queued to an unbound wq are served by the special
195 Work items of a highpri wq are queued to the highpri
243 achieve this behavior. Work items on such wq were always queued to the
[all …]
/Linux-v6.1/net/ax25/
Dax25_std_in.c143 int queued = 0; in ax25_std_state3_machine() local
225 queued = ax25_rx_iframe(ax25, skb); in ax25_std_state3_machine()
258 return queued; in ax25_std_state3_machine()
268 int queued = 0; in ax25_std_state4_machine() local
380 queued = ax25_rx_iframe(ax25, skb); in ax25_std_state4_machine()
413 return queued; in ax25_std_state4_machine()
421 int queued = 0, frametype, ns, nr, pf; in ax25_std_frame_in() local
427 queued = ax25_std_state1_machine(ax25, skb, frametype, pf, type); in ax25_std_frame_in()
430 queued = ax25_std_state2_machine(ax25, skb, frametype, pf, type); in ax25_std_frame_in()
433 queued = ax25_std_state3_machine(ax25, skb, frametype, ns, nr, pf, type); in ax25_std_frame_in()
[all …]
Dax25_ds_in.c147 int queued = 0; in ax25_ds_state3_machine() local
240 queued = ax25_rx_iframe(ax25, skb); in ax25_ds_state3_machine()
273 return queued; in ax25_ds_state3_machine()
281 int queued = 0, frametype, ns, nr, pf; in ax25_ds_frame_in() local
287 queued = ax25_ds_state1_machine(ax25, skb, frametype, pf, type); in ax25_ds_frame_in()
290 queued = ax25_ds_state2_machine(ax25, skb, frametype, pf, type); in ax25_ds_frame_in()
293 queued = ax25_ds_state3_machine(ax25, skb, frametype, ns, nr, pf, type); in ax25_ds_frame_in()
297 return queued; in ax25_ds_frame_in()
Dax25_in.c103 int queued = 0; in ax25_rx_iframe() local
145 queued = 1; in ax25_rx_iframe()
151 return queued; in ax25_rx_iframe()
159 int queued = 0; in ax25_process_rx_frame() local
167 queued = ax25_std_frame_in(ax25, skb, type); in ax25_process_rx_frame()
173 queued = ax25_ds_frame_in(ax25, skb, type); in ax25_process_rx_frame()
175 queued = ax25_std_frame_in(ax25, skb, type); in ax25_process_rx_frame()
180 return queued; in ax25_process_rx_frame()
305 * Process the frame. If it is queued up internally it in ax25_rcv()
/Linux-v6.1/include/drm/
Ddrm_flip_work.h47 * @val: value queued via drm_flip_work_queue()
69 * @queued: queued tasks
71 * @lock: lock to access queued and commited lists
77 struct list_head queued; member
/Linux-v6.1/kernel/locking/
Dqrwlock.c3 * Queued read/write locks
18 * queued_read_lock_slowpath - acquire read lock of a queued rwlock
19 * @lock: Pointer to queued rwlock structure
63 * queued_write_lock_slowpath - acquire write lock of a queued rwlock
64 * @lock : Pointer to queued rwlock structure
Dqspinlock.c3 * Queued spinlock
28 * Include queued spinlock statistics code
41 * This queued spinlock implementation is based on the MCS lock, however to
145 * @lock: Pointer to queued spinlock structure
156 * @lock: Pointer to queued spinlock structure
169 * @lock : Pointer to queued spinlock structure
191 * @lock: Pointer to queued spinlock structure
202 * @lock: Pointer to queued spinlock structure
213 * @lock : Pointer to queued spinlock structure
244 * @lock : Pointer to queued spinlock structure
[all …]
/Linux-v6.1/drivers/media/platform/renesas/vsp1/
Dvsp1_dl.c208 * @lock: protects the free, active, queued, and pending lists
211 * @queued: list queued to the hardware (written to the DL registers)
212 * @pending: list waiting to be queued to the hardware
224 struct vsp1_dl_list *queued; member
791 * loop on the same list until a new one is queued. In singleshot mode in vsp1_dl_list_fill_header()
841 if (!dlm->queued) in vsp1_dl_list_hw_update_pending()
871 * If a previous display list has been queued to the hardware but not in vsp1_dl_list_commit_continuous()
873 * case we can't replace the queued list by the new one, as we could in vsp1_dl_list_commit_continuous()
875 * be queued up to the hardware by the frame end interrupt handler. in vsp1_dl_list_commit_continuous()
894 * Pass the new display list to the hardware and mark it as queued. It in vsp1_dl_list_commit_continuous()
[all …]
/Linux-v6.1/Documentation/userspace-api/media/v4l/
Dvidioc-streamon.rst51 If ``VIDIOC_STREAMON`` fails then any already queued buffers will remain
52 queued.
63 If buffers have been queued with :ref:`VIDIOC_QBUF` and
65 ``VIDIOC_STREAMON``, then those queued buffers will also be removed from
77 but ``VIDIOC_STREAMOFF`` will return queued buffers to their starting
Ddev-encoder.rst309 fast raw frames are queued on the ``OUTPUT`` queue.
468 * a buffer queued to ``OUTPUT`` may result in more than one buffer produced on
473 * a buffer queued to ``OUTPUT`` may result in a buffer being produced on
479 buffers queued to ``OUTPUT`` (e.g. during drain or ``EOS``), because of the
480 ``OUTPUT`` buffers queued in the past whose encoding results are only
483 * buffers queued to ``OUTPUT`` may not become available to dequeue instantly
549 sequence to avoid losing the already queued/encoded frames.
559 To ensure that all the queued ``OUTPUT`` buffers have been processed and the
562 received all encoded frames for all ``OUTPUT`` buffers queued before the
586 2. Any ``OUTPUT`` buffers queued by the client before the
[all …]
/Linux-v6.1/net/dccp/
Dinput.c45 int queued = 0; in dccp_rcv_close() local
76 queued = 1; in dccp_rcv_close()
86 return queued; in dccp_rcv_close()
91 int queued = 0; in dccp_rcv_closereq() local
101 return queued; in dccp_rcv_closereq()
113 queued = 1; in dccp_rcv_closereq()
120 return queued; in dccp_rcv_closereq()
524 int queued = 0; in dccp_rcv_respond_partopen_state_process() local
562 queued = 1; /* packet was queued in dccp_rcv_respond_partopen_state_process()
568 return queued; in dccp_rcv_respond_partopen_state_process()
[all …]
/Linux-v6.1/net/netrom/
Dnr_in.c153 int queued = 0; in nr_state3_machine() local
225 queued = 1; in nr_state3_machine()
272 return queued; in nr_state3_machine()
279 int queued = 0, frametype; in nr_process_rx_frame() local
288 queued = nr_state1_machine(sk, skb, frametype); in nr_process_rx_frame()
291 queued = nr_state2_machine(sk, skb, frametype); in nr_process_rx_frame()
294 queued = nr_state3_machine(sk, skb, frametype); in nr_process_rx_frame()
300 return queued; in nr_process_rx_frame()
/Linux-v6.1/tools/testing/selftests/net/mptcp/
Dmptcp_inq.c207 int nsd, ret, queued = -1; in wait_for_ack() local
210 ret = ioctl(fd, TIOCOUTQ, &queued); in wait_for_ack()
218 if ((size_t)queued > total) in wait_for_ack()
219 xerror("TIOCOUTQ %u, but only %zu expected\n", queued, total); in wait_for_ack()
220 assert(nsd <= queued); in wait_for_ack()
222 if (queued == 0) in wait_for_ack()
231 xerror("still tx data queued after %u ms\n", timeout); in wait_for_ack()
355 unsigned int queued; in process_one_client() local
357 ret = ioctl(fd, FIONREAD, &queued); in process_one_client()
360 if (queued > expect_len) in process_one_client()
[all …]
/Linux-v6.1/Documentation/networking/
Drds.rst82 a socket. A message is queued when sendmsg is called, and
108 of queued bytes over the SO_SNDSIZE threshold will return
115 Receives a message that was queued to this socket. The sockets
132 incoming message queued to the socket, or a pending notification,
137 there's room on the send queue (ie the number of bytes queued
148 This allows the application to discard all messages queued to a
253 number of bytes queued equals or exceeds rcvbuf then the socket
296 This is then queued for the individual connection and sent by the
301 the transport to the general code and queued by the general code
331 Dropping a connection while packets are queued will cause queued or

12345678910>>...59