/Linux-v4.19/drivers/media/pci/saa7164/ |
D | saa7164-cmd.c | 32 ret = dev->cmds[i].seqno; in saa7164_cmd_alloc_seqno() 41 static void saa7164_cmd_free_seqno(struct saa7164_dev *dev, u8 seqno) in saa7164_cmd_free_seqno() argument 44 if ((dev->cmds[seqno].inuse == 1) && in saa7164_cmd_free_seqno() 45 (dev->cmds[seqno].seqno == seqno)) { in saa7164_cmd_free_seqno() 46 dev->cmds[seqno].inuse = 0; in saa7164_cmd_free_seqno() 47 dev->cmds[seqno].signalled = 0; in saa7164_cmd_free_seqno() 48 dev->cmds[seqno].timeout = 0; in saa7164_cmd_free_seqno() 53 static void saa7164_cmd_timeout_seqno(struct saa7164_dev *dev, u8 seqno) in saa7164_cmd_timeout_seqno() argument 56 if ((dev->cmds[seqno].inuse == 1) && in saa7164_cmd_timeout_seqno() 57 (dev->cmds[seqno].seqno == seqno)) { in saa7164_cmd_timeout_seqno() [all …]
|
/Linux-v4.19/drivers/gpu/drm/vc4/ |
D | vc4_trace.h | 21 TP_PROTO(struct drm_device *dev, uint64_t seqno, uint64_t timeout), 22 TP_ARGS(dev, seqno, timeout), 26 __field(u64, seqno) 32 __entry->seqno = seqno; 37 __entry->dev, __entry->seqno, __entry->timeout) 41 TP_PROTO(struct drm_device *dev, uint64_t seqno), 42 TP_ARGS(dev, seqno), 46 __field(u64, seqno) 51 __entry->seqno = seqno; 55 __entry->dev, __entry->seqno)
|
D | vc4_gem.c | 374 vc4_wait_for_seqno(struct drm_device *dev, uint64_t seqno, uint64_t timeout_ns, in vc4_wait_for_seqno() argument 382 if (vc4->finished_seqno >= seqno) in vc4_wait_for_seqno() 390 trace_vc4_wait_for_seqno_begin(dev, seqno, timeout_ns); in vc4_wait_for_seqno() 401 if (vc4->finished_seqno >= seqno) in vc4_wait_for_seqno() 416 trace_vc4_wait_for_seqno_end(dev, seqno); in vc4_wait_for_seqno() 530 vc4_update_bo_seqnos(struct vc4_exec_info *exec, uint64_t seqno) in vc4_update_bo_seqnos() argument 537 bo->seqno = seqno; in vc4_update_bo_seqnos() 543 bo->seqno = seqno; in vc4_update_bo_seqnos() 548 bo->write_seqno = seqno; in vc4_update_bo_seqnos() 664 uint64_t seqno; in vc4_queue_submit() local [all …]
|
/Linux-v4.19/drivers/gpu/drm/vmwgfx/ |
D | vmwgfx_irq.c | 110 static bool vmw_fifo_idle(struct vmw_private *dev_priv, uint32_t seqno) in vmw_fifo_idle() argument 120 uint32_t seqno = vmw_mmio_read(fifo_mem + SVGA_FIFO_FENCE); in vmw_update_seqno() local 122 if (dev_priv->last_read_seqno != seqno) { in vmw_update_seqno() 123 dev_priv->last_read_seqno = seqno; in vmw_update_seqno() 124 vmw_marker_pull(&fifo_state->marker_queue, seqno); in vmw_update_seqno() 130 uint32_t seqno) in vmw_seqno_passed() argument 135 if (likely(dev_priv->last_read_seqno - seqno < VMW_FENCE_WRAP)) in vmw_seqno_passed() 140 if (likely(dev_priv->last_read_seqno - seqno < VMW_FENCE_WRAP)) in vmw_seqno_passed() 144 vmw_fifo_idle(dev_priv, seqno)) in vmw_seqno_passed() 152 ret = ((atomic_read(&dev_priv->marker_seq) - seqno) in vmw_seqno_passed() [all …]
|
D | vmwgfx_marker.c | 33 uint32_t seqno; member 57 uint32_t seqno) in vmw_marker_push() argument 64 marker->seqno = seqno; in vmw_marker_push() 91 if (signaled_seqno - marker->seqno > (1 << 30)) in vmw_marker_pull() 132 uint32_t seqno; in vmw_wait_lag() local 138 seqno = atomic_read(&dev_priv->marker_seq); in vmw_wait_lag() 142 seqno = marker->seqno; in vmw_wait_lag() 146 ret = vmw_wait_seqno(dev_priv, false, seqno, true, in vmw_wait_lag() 152 (void) vmw_marker_pull(queue, seqno); in vmw_wait_lag()
|
/Linux-v4.19/drivers/gpu/drm/i915/ |
D | intel_breadcrumbs.c | 260 wait->seqno)); in intel_engine_disarm_breadcrumbs() 396 u32 seqno; in __intel_engine_add_wait() local 398 GEM_BUG_ON(!wait->seqno); in __intel_engine_add_wait() 416 seqno = intel_engine_get_seqno(engine); in __intel_engine_add_wait() 423 if (i915_seqno_passed(seqno, wait->seqno)) { in __intel_engine_add_wait() 431 if (wait->seqno == to_wait(parent)->seqno) { in __intel_engine_add_wait() 443 } else if (i915_seqno_passed(wait->seqno, in __intel_engine_add_wait() 444 to_wait(parent)->seqno)) { in __intel_engine_add_wait() 446 if (i915_seqno_passed(seqno, to_wait(parent)->seqno)) in __intel_engine_add_wait() 512 wait->seqno - 1); in intel_engine_add_wait() [all …]
|
D | i915_request.c | 199 static int reset_all_global_seqno(struct drm_i915_private *i915, u32 seqno) in reset_all_global_seqno() argument 220 engine->timeline.seqno, in reset_all_global_seqno() 222 seqno); in reset_all_global_seqno() 224 if (!i915_seqno_passed(seqno, engine->timeline.seqno)) { in reset_all_global_seqno() 236 intel_engine_init_global_seqno(engine, seqno); in reset_all_global_seqno() 237 engine->timeline.seqno = seqno; in reset_all_global_seqno() 243 i915->gt.request_serial = seqno; in reset_all_global_seqno() 248 int i915_gem_set_global_seqno(struct drm_device *dev, u32 seqno) in i915_gem_set_global_seqno() argument 254 if (seqno == 0) in i915_gem_set_global_seqno() 258 return reset_all_global_seqno(i915, seqno - 1); in i915_gem_set_global_seqno() [all …]
|
D | i915_timeline.h | 37 u32 seqno; member 103 u64 context, u32 seqno) in __i915_timeline_sync_set() argument 105 return i915_syncmap_set(&tl->sync, context, seqno); in __i915_timeline_sync_set() 111 return __i915_timeline_sync_set(tl, fence->context, fence->seqno); in i915_timeline_sync_set() 115 u64 context, u32 seqno) in __i915_timeline_sync_is_later() argument 117 return i915_syncmap_is_later(&tl->sync, context, seqno); in __i915_timeline_sync_is_later() 123 return __i915_timeline_sync_is_later(tl, fence->context, fence->seqno); in i915_timeline_sync_is_later()
|
D | i915_request.h | 46 u32 seqno; member 286 __i915_request_completed(const struct i915_request *rq, u32 seqno) in __i915_request_completed() argument 288 GEM_BUG_ON(!seqno); in __i915_request_completed() 289 return i915_seqno_passed(intel_engine_get_seqno(rq->engine), seqno) && in __i915_request_completed() 290 seqno == i915_request_global_seqno(rq); in __i915_request_completed() 295 u32 seqno; in i915_request_completed() local 297 seqno = i915_request_global_seqno(rq); in i915_request_completed() 298 if (!seqno) in i915_request_completed() 301 return __i915_request_completed(rq, seqno); in i915_request_completed() 306 u32 seqno; in i915_request_started() local [all …]
|
D | i915_syncmap.c | 154 bool i915_syncmap_is_later(struct i915_syncmap **root, u64 id, u32 seqno) in i915_syncmap_is_later() argument 195 return seqno_later(__sync_seqno(p)[idx], seqno); in i915_syncmap_is_later() 214 static inline void __sync_set_seqno(struct i915_syncmap *p, u64 id, u32 seqno) in __sync_set_seqno() argument 219 __sync_seqno(p)[idx] = seqno; in __sync_set_seqno() 230 static noinline int __sync_set(struct i915_syncmap **root, u64 id, u32 seqno) in __sync_set() argument 335 __sync_set_seqno(p, id, seqno); in __sync_set() 353 int i915_syncmap_set(struct i915_syncmap **root, u64 id, u32 seqno) in i915_syncmap_set() argument 362 __sync_set_seqno(p, id, seqno); in i915_syncmap_set() 366 return __sync_set(root, id, seqno); in i915_syncmap_set()
|
D | intel_hangcheck.c | 59 semaphore_waits_for(struct intel_engine_cs *engine, u32 *seqno) in semaphore_waits_for() argument 122 *seqno = ioread32(vaddr + head + 4) + 1; in semaphore_waits_for() 130 u32 seqno; in semaphore_passed() local 134 signaller = semaphore_waits_for(engine, &seqno); in semaphore_passed() 145 if (i915_seqno_passed(intel_engine_get_seqno(signaller), seqno)) in semaphore_passed() 287 hc->seqno = intel_engine_get_seqno(engine); in hangcheck_load_sample() 294 engine->hangcheck.seqno = hc->seqno; in hangcheck_store_sample() 304 if (engine->hangcheck.seqno != hc->seqno) in hangcheck_get_action()
|
D | i915_trace.h | 598 __field(u32, seqno) 607 __entry->seqno = from->global_seqno; 614 __entry->seqno) 627 __field(u32, seqno) 637 __entry->seqno = rq->fence.seqno; 643 __entry->hw_id, __entry->ctx, __entry->seqno, 657 __field(u32, seqno) 667 __entry->seqno = rq->fence.seqno; 673 __entry->hw_id, __entry->ctx, __entry->seqno, 703 __field(u32, seqno) [all …]
|
/Linux-v4.19/drivers/gpu/drm/i915/selftests/ |
D | i915_syncmap.c | 146 static int check_seqno(struct i915_syncmap *leaf, unsigned int idx, u32 seqno) in check_seqno() argument 154 if (__sync_seqno(leaf)[idx] != seqno) { in check_seqno() 156 __func__, idx, __sync_seqno(leaf)[idx], seqno); in check_seqno() 163 static int check_one(struct i915_syncmap **sync, u64 context, u32 seqno) in check_one() argument 167 err = i915_syncmap_set(sync, context, seqno); in check_one() 189 err = check_seqno((*sync), ilog2((*sync)->bitmap), seqno); in check_one() 193 if (!i915_syncmap_is_later(sync, context, seqno)) { in check_one() 195 context, seqno); in check_one() 238 static int check_leaf(struct i915_syncmap **sync, u64 context, u32 seqno) in check_leaf() argument 242 err = i915_syncmap_set(sync, context, seqno); in check_leaf() [all …]
|
D | intel_breadcrumbs.c | 52 idx, w->seqno); in check_rbtree() 58 idx, w->seqno, n); in check_rbtree() 80 n, waiters[n].seqno, in check_completion() 214 int seqno = 2 * n; in igt_insert_complete() local 220 n, waiters[n].seqno); in igt_insert_complete() 226 mock_seqno_advance(engine, seqno + seqno_bias); in igt_insert_complete() 227 for (m = n; m <= seqno; m++) { in igt_insert_complete() 241 seqno + seqno_bias); in igt_insert_complete() 248 seqno + seqno_bias); in igt_insert_complete() 271 u32 seqno; member [all …]
|
D | i915_timeline.c | 15 u32 seqno; member 27 if (__i915_timeline_sync_is_later(tl, ctx, p->seqno) != p->expected) { in __igt_sync() 29 name, p->name, ctx, p->seqno, yesno(p->expected)); in __igt_sync() 34 ret = __i915_timeline_sync_set(tl, ctx, p->seqno); in __igt_sync() 212 u32 seqno = prandom_u32_state(&prng); in bench_sync() local 214 if (!__i915_timeline_sync_is_later(&tl, id, seqno)) in bench_sync() 215 __i915_timeline_sync_set(&tl, id, seqno); in bench_sync()
|
/Linux-v4.19/drivers/gpu/drm/radeon/ |
D | radeon_trace.h | 127 TP_PROTO(struct drm_device *dev, int ring, u32 seqno), 129 TP_ARGS(dev, ring, seqno), 134 __field(u32, seqno) 140 __entry->seqno = seqno; 144 __entry->dev, __entry->ring, __entry->seqno) 149 TP_PROTO(struct drm_device *dev, int ring, u32 seqno), 151 TP_ARGS(dev, ring, seqno) 156 TP_PROTO(struct drm_device *dev, int ring, u32 seqno), 158 TP_ARGS(dev, ring, seqno) 163 TP_PROTO(struct drm_device *dev, int ring, u32 seqno), [all …]
|
/Linux-v4.19/drivers/gpu/drm/v3d/ |
D | v3d_trace.h | 17 uint64_t seqno, 19 TP_ARGS(dev, is_render, seqno, ctnqba, ctnqea), 24 __field(u64, seqno) 32 __entry->seqno = seqno; 40 __entry->seqno,
|
/Linux-v4.19/net/dccp/ |
D | ackvec.c | 54 int dccp_ackvec_update_records(struct dccp_ackvec *av, u64 seqno, u8 nonce_sum) in dccp_ackvec_update_records() argument 62 avr->avr_ack_seqno = seqno; in dccp_ackvec_update_records() 135 u64 seqno, enum dccp_ackvec_states state) in dccp_ackvec_update_old() argument 166 (unsigned long long)seqno, state); in dccp_ackvec_update_old() 200 u64 seqno, enum dccp_ackvec_states state) in dccp_ackvec_add_new() argument 240 av->av_buf_ackno = seqno; in dccp_ackvec_add_new() 251 u64 seqno = DCCP_SKB_CB(skb)->dccpd_seq; in dccp_ackvec_input() local 255 dccp_ackvec_add_new(av, 1, seqno, state); in dccp_ackvec_input() 256 av->av_tail_ackno = seqno; in dccp_ackvec_input() 259 s64 num_packets = dccp_delta_seqno(av->av_buf_ackno, seqno); in dccp_ackvec_input() [all …]
|
/Linux-v4.19/drivers/net/wireless/mediatek/mt76/ |
D | agg-rx.c | 88 mt76_rx_aggr_release_frames(tid, frames, status->seqno); in mt76_rx_aggr_check_release() 129 u16 seqno; in mt76_rx_aggr_check_ctl() local 138 seqno = le16_to_cpu(bar->start_seq_num) >> 4; in mt76_rx_aggr_check_ctl() 144 mt76_rx_aggr_release_frames(tid, frames, seqno); in mt76_rx_aggr_check_ctl() 157 u16 seqno, head, size; in mt76_rx_aggr_reorder() local 188 seqno = status->seqno; in mt76_rx_aggr_reorder() 190 sn_less = ieee80211_sn_less(seqno, head); in mt76_rx_aggr_reorder() 205 if (seqno == head) { in mt76_rx_aggr_reorder() 218 if (!ieee80211_sn_less(seqno, head + size)) { in mt76_rx_aggr_reorder() 219 head = ieee80211_sn_inc(ieee80211_sn_sub(seqno, size)); in mt76_rx_aggr_reorder() [all …]
|
/Linux-v4.19/net/dccp/ccids/lib/ |
D | packet_history.h | 51 u64 seqno; member 56 tfrc_tx_hist_find_entry(struct tfrc_tx_hist_entry *head, u64 seqno) in tfrc_tx_hist_find_entry() argument 58 while (head != NULL && head->seqno != seqno) in tfrc_tx_hist_find_entry() 63 int tfrc_tx_hist_add(struct tfrc_tx_hist_entry **headp, u64 seqno);
|
/Linux-v4.19/include/uapi/linux/ |
D | batadv_packet.h | 217 __be32 seqno; member 243 __be32 seqno; member 263 __be32 seqno; member 316 __be16 seqno; member 346 __be32 seqno; member 384 __be16 seqno; member 467 __be16 seqno; member 485 __be32 seqno; member
|
/Linux-v4.19/include/linux/ |
D | dma-fence.h | 80 unsigned seqno; member 247 spinlock_t *lock, u64 context, unsigned seqno); 436 return __dma_fence_is_later(f1->seqno, f2->seqno); in dma_fence_is_later() 551 __ff->context, __ff->seqno, ##args); \ 557 pr_warn("f %llu#%u: " fmt, __ff->context, __ff->seqno, \ 564 pr_err("f %llu#%u: " fmt, __ff->context, __ff->seqno, \
|
/Linux-v4.19/net/batman-adv/ |
D | fragmentation.c | 121 u16 seqno) in batadv_frag_init_chain() argument 125 if (chain->seqno == seqno) in batadv_frag_init_chain() 132 chain->seqno = seqno; in batadv_frag_init_chain() 159 u16 seqno, hdr_size = sizeof(struct batadv_frag_packet); in batadv_frag_insert_packet() local 170 seqno = ntohs(frag_packet->seqno); in batadv_frag_insert_packet() 171 bucket = seqno % BATADV_FRAG_BUFFER_COUNT; in batadv_frag_insert_packet() 186 if (batadv_frag_init_chain(chain, seqno)) { in batadv_frag_insert_packet() 492 frag_header.seqno = htons(atomic_inc_return(&bat_priv->frag_seqno)); in batadv_frag_send_packet()
|
/Linux-v4.19/include/trace/events/ |
D | dma_fence.h | 22 __field(unsigned int, seqno) 29 __entry->seqno = fence->seqno; 34 __entry->seqno)
|
/Linux-v4.19/drivers/net/ppp/ |
D | ppp_deflate.c | 28 int seqno; member 153 state->seqno = 0; in z_comp_init() 173 state->seqno = 0; in z_comp_reset() 216 put_unaligned_be16(state->seqno, wptr); in z_compress() 221 ++state->seqno; in z_compress() 368 state->seqno = 0; in z_decomp_init() 389 state->seqno = 0; in z_decomp_reset() 431 if (seq != (state->seqno & 0xffff)) { in z_decompress() 434 state->unit, seq, state->seqno & 0xffff); in z_decompress() 437 ++state->seqno; in z_decompress() [all …]
|