Home
last modified time | relevance | path

Searched refs:seqno (Results 1 – 25 of 223) sorted by relevance

123456789

/kernel/linux/linux-5.10/drivers/gpu/drm/v3d/
Dv3d_trace.h39 uint64_t seqno,
41 TP_ARGS(dev, is_render, seqno, ctnqba, ctnqea),
46 __field(u64, seqno)
54 __entry->seqno = seqno;
62 __entry->seqno,
69 uint64_t seqno),
70 TP_ARGS(dev, seqno),
74 __field(u64, seqno)
79 __entry->seqno = seqno;
84 __entry->seqno)
[all …]
/kernel/linux/linux-5.10/drivers/media/pci/saa7164/
Dsaa7164-cmd.c22 ret = dev->cmds[i].seqno; in saa7164_cmd_alloc_seqno()
31 static void saa7164_cmd_free_seqno(struct saa7164_dev *dev, u8 seqno) in saa7164_cmd_free_seqno() argument
34 if ((dev->cmds[seqno].inuse == 1) && in saa7164_cmd_free_seqno()
35 (dev->cmds[seqno].seqno == seqno)) { in saa7164_cmd_free_seqno()
36 dev->cmds[seqno].inuse = 0; in saa7164_cmd_free_seqno()
37 dev->cmds[seqno].signalled = 0; in saa7164_cmd_free_seqno()
38 dev->cmds[seqno].timeout = 0; in saa7164_cmd_free_seqno()
43 static void saa7164_cmd_timeout_seqno(struct saa7164_dev *dev, u8 seqno) in saa7164_cmd_timeout_seqno() argument
46 if ((dev->cmds[seqno].inuse == 1) && in saa7164_cmd_timeout_seqno()
47 (dev->cmds[seqno].seqno == seqno)) { in saa7164_cmd_timeout_seqno()
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/vc4/
Dvc4_trace.h18 TP_PROTO(struct drm_device *dev, uint64_t seqno, uint64_t timeout),
19 TP_ARGS(dev, seqno, timeout),
23 __field(u64, seqno)
29 __entry->seqno = seqno;
34 __entry->dev, __entry->seqno, __entry->timeout)
38 TP_PROTO(struct drm_device *dev, uint64_t seqno),
39 TP_ARGS(dev, seqno),
43 __field(u64, seqno)
48 __entry->seqno = seqno;
52 __entry->dev, __entry->seqno)
Dvc4_gem.c381 vc4_wait_for_seqno(struct drm_device *dev, uint64_t seqno, uint64_t timeout_ns, in vc4_wait_for_seqno() argument
389 if (vc4->finished_seqno >= seqno) in vc4_wait_for_seqno()
397 trace_vc4_wait_for_seqno_begin(dev, seqno, timeout_ns); in vc4_wait_for_seqno()
408 if (vc4->finished_seqno >= seqno) in vc4_wait_for_seqno()
423 trace_vc4_wait_for_seqno_end(dev, seqno); in vc4_wait_for_seqno()
537 vc4_update_bo_seqnos(struct vc4_exec_info *exec, uint64_t seqno) in vc4_update_bo_seqnos() argument
544 bo->seqno = seqno; in vc4_update_bo_seqnos()
550 bo->seqno = seqno; in vc4_update_bo_seqnos()
555 bo->write_seqno = seqno; in vc4_update_bo_seqnos()
670 uint64_t seqno; in vc4_queue_submit() local
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/vmwgfx/
Dvmwgfx_irq.c111 static bool vmw_fifo_idle(struct vmw_private *dev_priv, uint32_t seqno) in vmw_fifo_idle() argument
121 uint32_t seqno = vmw_mmio_read(fifo_mem + SVGA_FIFO_FENCE); in vmw_update_seqno() local
123 if (dev_priv->last_read_seqno != seqno) { in vmw_update_seqno()
124 dev_priv->last_read_seqno = seqno; in vmw_update_seqno()
125 vmw_marker_pull(&fifo_state->marker_queue, seqno); in vmw_update_seqno()
131 uint32_t seqno) in vmw_seqno_passed() argument
136 if (likely(dev_priv->last_read_seqno - seqno < VMW_FENCE_WRAP)) in vmw_seqno_passed()
141 if (likely(dev_priv->last_read_seqno - seqno < VMW_FENCE_WRAP)) in vmw_seqno_passed()
145 vmw_fifo_idle(dev_priv, seqno)) in vmw_seqno_passed()
153 ret = ((atomic_read(&dev_priv->marker_seq) - seqno) in vmw_seqno_passed()
[all …]
Dvmwgfx_marker.c33 uint32_t seqno; member
57 uint32_t seqno) in vmw_marker_push() argument
64 marker->seqno = seqno; in vmw_marker_push()
91 if (signaled_seqno - marker->seqno > (1 << 30)) in vmw_marker_pull()
132 uint32_t seqno; in vmw_wait_lag() local
138 seqno = atomic_read(&dev_priv->marker_seq); in vmw_wait_lag()
142 seqno = marker->seqno; in vmw_wait_lag()
146 ret = vmw_wait_seqno(dev_priv, false, seqno, true, in vmw_wait_lag()
152 (void) vmw_marker_pull(queue, seqno); in vmw_wait_lag()
/kernel/linux/linux-5.10/drivers/dma-buf/
Ddma-fence-chain.c90 int dma_fence_chain_find_seqno(struct dma_fence **pfence, uint64_t seqno) in dma_fence_chain_find_seqno() argument
94 if (!seqno) in dma_fence_chain_find_seqno()
98 if (!chain || chain->base.seqno < seqno) in dma_fence_chain_find_seqno()
103 to_dma_fence_chain(*pfence)->prev_seqno < seqno) in dma_fence_chain_find_seqno()
233 uint64_t seqno) in dma_fence_chain_init() argument
245 if (prev_chain && __dma_fence_is_later(seqno, prev->seqno, prev->ops)) { in dma_fence_chain_init()
247 chain->prev_seqno = prev->seqno; in dma_fence_chain_init()
252 seqno = max(prev->seqno, seqno); in dma_fence_chain_init()
256 &chain->lock, context, seqno); in dma_fence_chain_init()
Dst-dma-fence-chain.c69 u64 seqno) in mock_chain() argument
80 seqno); in mock_chain()
285 fence->seqno); in find_signaled()
330 fence ? fence->seqno : 0); in find_out_of_order()
367 fence->seqno, in find_gap()
405 int seqno; in __find_race() local
407 seqno = prandom_u32_max(data->fc.chain_length) + 1; in __find_race()
409 err = dma_fence_chain_find_seqno(&fence, seqno); in __find_race()
412 seqno); in __find_race()
423 if (fence->seqno == seqno) { in __find_race()
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/selftests/
Di915_syncmap.c146 static int check_seqno(struct i915_syncmap *leaf, unsigned int idx, u32 seqno) in check_seqno() argument
154 if (__sync_seqno(leaf)[idx] != seqno) { in check_seqno()
156 __func__, idx, __sync_seqno(leaf)[idx], seqno); in check_seqno()
163 static int check_one(struct i915_syncmap **sync, u64 context, u32 seqno) in check_one() argument
167 err = i915_syncmap_set(sync, context, seqno); in check_one()
189 err = check_seqno((*sync), ilog2((*sync)->bitmap), seqno); in check_one()
193 if (!i915_syncmap_is_later(sync, context, seqno)) { in check_one()
195 context, seqno); in check_one()
238 static int check_leaf(struct i915_syncmap **sync, u64 context, u32 seqno) in check_leaf() argument
242 err = i915_syncmap_set(sync, context, seqno); in check_leaf()
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/
Dintel_timeline.h68 u64 context, u32 seqno) in __intel_timeline_sync_set() argument
70 return i915_syncmap_set(&tl->sync, context, seqno); in __intel_timeline_sync_set()
76 return __intel_timeline_sync_set(tl, fence->context, fence->seqno); in intel_timeline_sync_set()
80 u64 context, u32 seqno) in __intel_timeline_sync_is_later() argument
82 return i915_syncmap_is_later(&tl->sync, context, seqno); in __intel_timeline_sync_is_later()
88 return __intel_timeline_sync_is_later(tl, fence->context, fence->seqno); in intel_timeline_sync_is_later()
96 u32 *seqno);
Dselftest_timeline.c167 u32 seqno; member
179 if (__intel_timeline_sync_is_later(tl, ctx, p->seqno) != p->expected) { in __igt_sync()
181 name, p->name, ctx, p->seqno, yesno(p->expected)); in __igt_sync()
186 ret = __intel_timeline_sync_set(tl, ctx, p->seqno); in __igt_sync()
364 u32 seqno = prandom_u32_state(&prng); in bench_sync() local
366 if (!__intel_timeline_sync_is_later(&tl, id, seqno)) in bench_sync()
367 __intel_timeline_sync_set(&tl, id, seqno); in bench_sync()
494 if (READ_ONCE(*tl->hwsp_seqno) != tl->seqno) { in checked_intel_timeline_create()
496 *tl->hwsp_seqno, tl->seqno); in checked_intel_timeline_create()
677 u32 seqno[2]; in live_hwsp_wrap() local
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/radeon/
Dradeon_trace.h127 TP_PROTO(struct drm_device *dev, int ring, u32 seqno),
129 TP_ARGS(dev, ring, seqno),
134 __field(u32, seqno)
140 __entry->seqno = seqno;
144 __entry->dev, __entry->ring, __entry->seqno)
149 TP_PROTO(struct drm_device *dev, int ring, u32 seqno),
151 TP_ARGS(dev, ring, seqno)
156 TP_PROTO(struct drm_device *dev, int ring, u32 seqno),
158 TP_ARGS(dev, ring, seqno)
163 TP_PROTO(struct drm_device *dev, int ring, u32 seqno),
[all …]
/kernel/linux/linux-5.10/include/trace/events/
Drpcgss.h298 __field(u32, seqno)
305 __entry->seqno = gc->gc_seq;
310 __entry->xid, __entry->seqno)
377 __field(u32, seqno)
386 __entry->seqno = rqst->rq_seqno;
391 __entry->xid, __entry->seqno)
408 __field(u32, seqno)
417 __entry->seqno = task->tk_rqstp->rq_seqno;
423 __entry->xid, __entry->seqno, __entry->seq_xmit,
464 u32 seqno
[all …]
/kernel/linux/linux-5.10/net/dccp/
Dackvec.c51 int dccp_ackvec_update_records(struct dccp_ackvec *av, u64 seqno, u8 nonce_sum) in dccp_ackvec_update_records() argument
59 avr->avr_ack_seqno = seqno; in dccp_ackvec_update_records()
132 u64 seqno, enum dccp_ackvec_states state) in dccp_ackvec_update_old() argument
163 (unsigned long long)seqno, state); in dccp_ackvec_update_old()
197 u64 seqno, enum dccp_ackvec_states state) in dccp_ackvec_add_new() argument
237 av->av_buf_ackno = seqno; in dccp_ackvec_add_new()
248 u64 seqno = DCCP_SKB_CB(skb)->dccpd_seq; in dccp_ackvec_input() local
252 dccp_ackvec_add_new(av, 1, seqno, state); in dccp_ackvec_input()
253 av->av_tail_ackno = seqno; in dccp_ackvec_input()
256 s64 num_packets = dccp_delta_seqno(av->av_buf_ackno, seqno); in dccp_ackvec_input()
[all …]
/kernel/linux/linux-5.10/drivers/net/wireless/mediatek/mt76/
Dagg-rx.c84 mt76_rx_aggr_release_frames(tid, frames, status->seqno); in mt76_rx_aggr_check_release()
125 u16 seqno; in mt76_rx_aggr_check_ctl() local
134 seqno = IEEE80211_SEQ_TO_SN(le16_to_cpu(bar->start_seq_num)); in mt76_rx_aggr_check_ctl()
141 mt76_rx_aggr_release_frames(tid, frames, seqno); in mt76_rx_aggr_check_ctl()
155 u16 seqno, head, size, idx; in mt76_rx_aggr_reorder() local
186 seqno = status->seqno; in mt76_rx_aggr_reorder()
188 sn_less = ieee80211_sn_less(seqno, head); in mt76_rx_aggr_reorder()
203 if (seqno == head) { in mt76_rx_aggr_reorder()
216 if (!ieee80211_sn_less(seqno, head + size)) { in mt76_rx_aggr_reorder()
217 head = ieee80211_sn_inc(ieee80211_sn_sub(seqno, size)); in mt76_rx_aggr_reorder()
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/virtio/
Dvirtgpu_fence.c47 if (WARN_ON_ONCE(fence->f.seqno == 0)) in virtio_fence_signaled()
51 if (atomic64_read(&fence->drv->last_seq) >= fence->f.seqno) in virtio_fence_signaled()
58 snprintf(str, size, "%llu", f->seqno); in virtio_fence_value_str()
103 fence->f.seqno = ++drv->sync_seq; in virtio_gpu_fence_emit()
111 cmd_hdr->fence_id = cpu_to_le64(fence->f.seqno); in virtio_gpu_fence_emit()
124 if (last_seq < fence->f.seqno) in virtio_gpu_fence_event_process()
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/
Di915_syncmap.c154 bool i915_syncmap_is_later(struct i915_syncmap **root, u64 id, u32 seqno) in i915_syncmap_is_later() argument
195 return seqno_later(__sync_seqno(p)[idx], seqno); in i915_syncmap_is_later()
214 static inline void __sync_set_seqno(struct i915_syncmap *p, u64 id, u32 seqno) in __sync_set_seqno() argument
219 __sync_seqno(p)[idx] = seqno; in __sync_set_seqno()
230 static noinline int __sync_set(struct i915_syncmap **root, u64 id, u32 seqno) in __sync_set() argument
335 __sync_set_seqno(p, id, seqno); in __sync_set()
353 int i915_syncmap_set(struct i915_syncmap **root, u64 id, u32 seqno) in i915_syncmap_set() argument
362 __sync_set_seqno(p, id, seqno); in i915_syncmap_set()
366 return __sync_set(root, id, seqno); in i915_syncmap_set()
/kernel/linux/linux-5.10/drivers/gpu/drm/msm/
Dmsm_gpu_trace.h40 __field(u32, seqno)
47 __entry->seqno = submit->seqno;
51 __entry->id, __entry->pid, __entry->ringid, __entry->seqno,
64 __field(u32, seqno)
74 __entry->seqno = submit->seqno;
81 __entry->id, __entry->pid, __entry->ringid, __entry->seqno,
/kernel/linux/linux-5.10/net/dccp/ccids/lib/
Dpacket_history.h38 u64 seqno; member
43 tfrc_tx_hist_find_entry(struct tfrc_tx_hist_entry *head, u64 seqno) in tfrc_tx_hist_find_entry() argument
45 while (head != NULL && head->seqno != seqno) in tfrc_tx_hist_find_entry()
50 int tfrc_tx_hist_add(struct tfrc_tx_hist_entry **headp, u64 seqno);
/kernel/linux/patches/linux-5.10/prebuilts/usr/include/linux/
Dbatadv_packet.h102 __be32 seqno; member
115 __be32 seqno; member
125 __be32 seqno; member
148 __be16 seqno; member
160 __be32 seqno; member
177 __be16 seqno; member
211 __be16 seqno; member
219 __be32 seqno; member
/kernel/linux/patches/linux-4.19/prebuilts/usr/include/linux/
Dbatadv_packet.h90 __be32 seqno; member
103 __be32 seqno; member
113 __be32 seqno; member
136 __be16 seqno; member
148 __be32 seqno; member
165 __be16 seqno; member
201 __be16 seqno; member
209 __be32 seqno; member
/kernel/linux/linux-5.10/include/uapi/linux/
Dbatadv_packet.h213 __be32 seqno; member
239 __be32 seqno; member
259 __be32 seqno; member
312 __be16 seqno; member
342 __be32 seqno; member
380 __be16 seqno; member
463 __be16 seqno; member
481 __be32 seqno; member
/kernel/linux/linux-5.10/drivers/gpu/drm/lima/
Dlima_trace.h19 __field(unsigned int, seqno)
26 __entry->seqno = task->base.s_fence->finished.seqno;
31 __entry->task_id, __entry->context, __entry->seqno,
/kernel/linux/linux-5.10/include/linux/
Ddma-fence.h92 u64 seqno; member
267 spinlock_t *lock, u64 context, u64 seqno);
478 return __dma_fence_is_later(f1->seqno, f2->seqno, f1->ops); in dma_fence_is_later()
594 __ff->context, __ff->seqno, ##args); \
600 pr_warn("f %llu#%llu: " fmt, __ff->context, __ff->seqno,\
607 pr_err("f %llu#%llu: " fmt, __ff->context, __ff->seqno, \
/kernel/linux/linux-5.10/net/batman-adv/
Dfragmentation.c109 u16 seqno) in batadv_frag_init_chain() argument
113 if (chain->seqno == seqno) in batadv_frag_init_chain()
120 chain->seqno = seqno; in batadv_frag_init_chain()
147 u16 seqno, hdr_size = sizeof(struct batadv_frag_packet); in batadv_frag_insert_packet() local
158 seqno = ntohs(frag_packet->seqno); in batadv_frag_insert_packet()
159 bucket = seqno % BATADV_FRAG_BUFFER_COUNT; in batadv_frag_insert_packet()
174 if (batadv_frag_init_chain(chain, seqno)) { in batadv_frag_insert_packet()
485 frag_header.seqno = htons(atomic_inc_return(&bat_priv->frag_seqno)); in batadv_frag_send_packet()

123456789