Home
last modified time | relevance | path

Searched refs:tl (Results 1 – 25 of 100) sorted by relevance

1234

/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/
Dselftest_timeline.c24 static struct page *hwsp_page(struct intel_timeline *tl) in hwsp_page() argument
26 struct drm_i915_gem_object *obj = tl->hwsp_ggtt->obj; in hwsp_page()
32 static unsigned long hwsp_cacheline(struct intel_timeline *tl) in hwsp_cacheline() argument
34 unsigned long address = (unsigned long)page_address(hwsp_page(tl)); in hwsp_cacheline()
36 return (address + tl->hwsp_offset) / CACHELINE_BYTES; in hwsp_cacheline()
55 struct intel_timeline *tl) in __mock_hwsp_record() argument
57 tl = xchg(&state->history[idx], tl); in __mock_hwsp_record()
58 if (tl) { in __mock_hwsp_record()
59 radix_tree_delete(&state->cachelines, hwsp_cacheline(tl)); in __mock_hwsp_record()
60 intel_timeline_put(tl); in __mock_hwsp_record()
[all …]
Dintel_gt_requests.c17 static bool retire_requests(struct intel_timeline *tl) in retire_requests() argument
21 list_for_each_entry_safe(rq, rn, &tl->requests, link) in retire_requests()
26 return !i915_active_fence_isset(&tl->last_request); in retire_requests()
64 struct intel_timeline *tl = xchg(&engine->retire, NULL); in engine_retire() local
67 struct intel_timeline *next = xchg(&tl->retire, NULL); in engine_retire()
77 if (mutex_trylock(&tl->mutex)) { in engine_retire()
78 retire_requests(tl); in engine_retire()
79 mutex_unlock(&tl->mutex); in engine_retire()
81 intel_timeline_put(tl); in engine_retire()
84 tl = ptr_mask_bits(next, 1); in engine_retire()
[all …]
Dintel_timeline.c330 void __intel_timeline_pin(struct intel_timeline *tl) in __intel_timeline_pin() argument
332 GEM_BUG_ON(!atomic_read(&tl->pin_count)); in __intel_timeline_pin()
333 atomic_inc(&tl->pin_count); in __intel_timeline_pin()
336 int intel_timeline_pin(struct intel_timeline *tl, struct i915_gem_ww_ctx *ww) in intel_timeline_pin() argument
340 if (atomic_add_unless(&tl->pin_count, 1, 0)) in intel_timeline_pin()
343 err = i915_ggtt_pin(tl->hwsp_ggtt, ww, 0, PIN_HIGH); in intel_timeline_pin()
347 tl->hwsp_offset = in intel_timeline_pin()
348 i915_ggtt_offset(tl->hwsp_ggtt) + in intel_timeline_pin()
349 offset_in_page(tl->hwsp_offset); in intel_timeline_pin()
350 GT_TRACE(tl->gt, "timeline:%llx using HWSP offset:%x\n", in intel_timeline_pin()
[all …]
Dintel_timeline.h67 static inline int __intel_timeline_sync_set(struct intel_timeline *tl, in __intel_timeline_sync_set() argument
70 return i915_syncmap_set(&tl->sync, context, seqno); in __intel_timeline_sync_set()
73 static inline int intel_timeline_sync_set(struct intel_timeline *tl, in intel_timeline_sync_set() argument
76 return __intel_timeline_sync_set(tl, fence->context, fence->seqno); in intel_timeline_sync_set()
79 static inline bool __intel_timeline_sync_is_later(struct intel_timeline *tl, in __intel_timeline_sync_is_later() argument
82 return i915_syncmap_is_later(&tl->sync, context, seqno); in __intel_timeline_sync_is_later()
85 static inline bool intel_timeline_sync_is_later(struct intel_timeline *tl, in intel_timeline_sync_is_later() argument
88 return __intel_timeline_sync_is_later(tl, fence->context, fence->seqno); in intel_timeline_sync_is_later()
91 void __intel_timeline_pin(struct intel_timeline *tl);
92 int intel_timeline_pin(struct intel_timeline *tl, struct i915_gem_ww_ctx *ww);
[all …]
Dselftest_context.c18 struct intel_timeline *tl = i915_request_timeline(rq); in request_sync() local
22 intel_timeline_get(tl); in request_sync()
36 lockdep_unpin_lock(&tl->mutex, rq->cookie); in request_sync()
37 mutex_unlock(&tl->mutex); in request_sync()
40 intel_timeline_put(tl); in request_sync()
47 struct intel_timeline *tl = ce->timeline; in context_sync() local
50 mutex_lock(&tl->mutex); in context_sync()
55 if (list_empty(&tl->requests)) in context_sync()
58 rq = list_last_entry(&tl->requests, typeof(*rq), link); in context_sync()
69 mutex_unlock(&tl->mutex); in context_sync()
Dintel_context.h158 struct intel_timeline *tl = ce->timeline; in intel_context_timeline_lock() local
161 err = mutex_lock_interruptible(&tl->mutex); in intel_context_timeline_lock()
165 return tl; in intel_context_timeline_lock()
168 static inline void intel_context_timeline_unlock(struct intel_timeline *tl) in intel_context_timeline_unlock() argument
169 __releases(&tl->mutex) in intel_context_timeline_unlock()
171 mutex_unlock(&tl->mutex); in intel_context_timeline_unlock()
Dmock_engine.c35 static void mock_timeline_pin(struct intel_timeline *tl) in mock_timeline_pin() argument
37 atomic_inc(&tl->pin_count); in mock_timeline_pin()
40 static void mock_timeline_unpin(struct intel_timeline *tl) in mock_timeline_unpin() argument
42 GEM_BUG_ON(!atomic_read(&tl->pin_count)); in mock_timeline_unpin()
43 atomic_dec(&tl->pin_count); in mock_timeline_unpin()
/kernel/linux/linux-5.10/fs/cifs/
Ddfs_cache.h56 dfs_cache_get_next_tgt(struct dfs_cache_tgt_list *tl, in dfs_cache_get_next_tgt() argument
59 if (!tl || list_empty(&tl->tl_list) || !it || in dfs_cache_get_next_tgt()
60 list_is_last(&it->it_list, &tl->tl_list)) in dfs_cache_get_next_tgt()
66 dfs_cache_get_tgt_iterator(struct dfs_cache_tgt_list *tl) in dfs_cache_get_tgt_iterator() argument
68 if (!tl) in dfs_cache_get_tgt_iterator()
70 return list_first_entry_or_null(&tl->tl_list, in dfs_cache_get_tgt_iterator()
75 static inline void dfs_cache_free_tgts(struct dfs_cache_tgt_list *tl) in dfs_cache_free_tgts() argument
79 if (!tl || list_empty(&tl->tl_list)) in dfs_cache_free_tgts()
81 list_for_each_entry_safe(it, nit, &tl->tl_list, it_list) { in dfs_cache_free_tgts()
86 tl->tl_numtgts = 0; in dfs_cache_free_tgts()
[all …]
/kernel/linux/linux-5.10/drivers/net/ethernet/netronome/nfp/
Dnfp_net_debugdump.c58 struct nfp_dump_tl tl; member
64 struct nfp_dump_tl tl; member
70 struct nfp_dump_tl tl; member
78 struct nfp_dump_tl tl; member
87 struct nfp_dump_tl tl; member
92 struct nfp_dump_tl tl; member
112 typedef int (*nfp_tlv_visit)(struct nfp_pf *pf, struct nfp_dump_tl *tl,
120 struct nfp_dump_tl *tl; in nfp_traverse_tlvs() local
125 while (remaining >= sizeof(*tl)) { in nfp_traverse_tlvs()
126 tl = p; in nfp_traverse_tlvs()
[all …]
/kernel/linux/linux-5.10/fs/ext4/
Dfast_commit.c602 struct ext4_fc_tl *tl; in ext4_fc_reserve_space() local
631 tl = (struct ext4_fc_tl *)(sbi->s_fc_bh->b_data + off); in ext4_fc_reserve_space()
632 tl->fc_tag = cpu_to_le16(EXT4_FC_TAG_PAD); in ext4_fc_reserve_space()
634 tl->fc_len = cpu_to_le16(pad_len); in ext4_fc_reserve_space()
636 *crc = ext4_chksum(sbi, *crc, tl, sizeof(*tl)); in ext4_fc_reserve_space()
638 ext4_fc_memzero(sb, tl + 1, pad_len, crc); in ext4_fc_reserve_space()
640 *((u8 *)(tl + 1) + pad_len) = 0; in ext4_fc_reserve_space()
672 struct ext4_fc_tl tl; in ext4_fc_write_tail() local
681 dst = ext4_fc_reserve_space(sb, sizeof(tl) + sizeof(tail), &crc); in ext4_fc_write_tail()
687 tl.fc_tag = cpu_to_le16(EXT4_FC_TAG_TAIL); in ext4_fc_write_tail()
[all …]
/kernel/linux/linux-5.10/crypto/
Dvmac.c150 int i; u64 th, tl; \
153 MUL64(th, tl, pe64_to_cpup((mp)+i)+(kp)[i], \
155 ADD128(rh, rl, th, tl); \
161 int i; u64 th, tl; \
164 MUL64(th, tl, pe64_to_cpup((mp)+i)+(kp)[i], \
166 ADD128(rh, rl, th, tl); \
167 MUL64(th, tl, pe64_to_cpup((mp)+i)+(kp)[i+2], \
169 ADD128(rh1, rl1, th, tl); \
176 int i; u64 th, tl; \
179 MUL64(th, tl, pe64_to_cpup((mp)+i)+(kp)[i], \
[all …]
Dcamellia_generic.c367 u32 dw, tl, tr; in camellia_setup_tail() local
466 tl = subL[10] ^ (subR[10] & ~subR[8]); in camellia_setup_tail()
467 dw = tl & subL[8]; /* FL(kl1) */ in camellia_setup_tail()
469 SUBKEY_L(7) = subL[6] ^ tl; /* round 6 */ in camellia_setup_tail()
475 tl = subL[7] ^ (subR[7] & ~subR[9]); in camellia_setup_tail()
476 dw = tl & subL[9]; /* FLinv(kl2) */ in camellia_setup_tail()
478 SUBKEY_L(10) = tl ^ subL[11]; /* round 7 */ in camellia_setup_tail()
488 tl = subL[18] ^ (subR[18] & ~subR[16]); in camellia_setup_tail()
489 dw = tl & subL[16]; /* FL(kl3) */ in camellia_setup_tail()
491 SUBKEY_L(15) = subL[14] ^ tl; /* round 12 */ in camellia_setup_tail()
[all …]
/kernel/linux/linux-5.10/drivers/isdn/mISDN/
Dfsm.c98 struct FsmTimer *ft = from_timer(ft, t, tl); in FsmExpireTimer()
114 timer_setup(&ft->tl, FsmExpireTimer, 0); in mISDN_FsmInitTimer()
126 del_timer(&ft->tl); in mISDN_FsmDelTimer()
141 if (timer_pending(&ft->tl)) { in mISDN_FsmAddTimer()
152 ft->tl.expires = jiffies + (millisec * HZ) / 1000; in mISDN_FsmAddTimer()
153 add_timer(&ft->tl); in mISDN_FsmAddTimer()
169 if (timer_pending(&ft->tl)) in mISDN_FsmRestartTimer()
170 del_timer(&ft->tl); in mISDN_FsmRestartTimer()
173 ft->tl.expires = jiffies + (millisec * HZ) / 1000; in mISDN_FsmRestartTimer()
174 add_timer(&ft->tl); in mISDN_FsmRestartTimer()
Dtimerdev.c39 struct timer_list tl; member
77 del_timer_sync(&timer->tl); in mISDN_close()
158 struct mISDNtimer *timer = from_timer(timer, t, tl); in dev_expire_timer()
183 timer_setup(&timer->tl, dev_expire_timer, 0); in misdn_add_timer()
189 timer->tl.expires = jiffies + ((HZ * (u_long)timeout) / 1000); in misdn_add_timer()
190 add_timer(&timer->tl); in misdn_add_timer()
207 del_timer_sync(&timer->tl); in misdn_del_timer()
Ddsp_tones.c462 struct dsp *dsp = from_timer(dsp, t, tone.tl); in dsp_tone_timeout()
481 tone->tl.expires = jiffies + (pat->seq[index] * HZ) / 8000; in dsp_tone_timeout()
482 add_timer(&tone->tl); in dsp_tone_timeout()
507 if (dsp->features.hfc_loops && timer_pending(&tonet->tl)) in dsp_tone()
508 del_timer(&tonet->tl); in dsp_tone()
541 if (timer_pending(&tonet->tl)) in dsp_tone()
542 del_timer(&tonet->tl); in dsp_tone()
543 tonet->tl.expires = jiffies + (pat->seq[0] * HZ) / 8000; in dsp_tone()
544 add_timer(&tonet->tl); in dsp_tone()
/kernel/linux/linux-5.10/drivers/s390/net/
Dfsm.c135 fsm_timer *this = from_timer(this, t, tl); in fsm_expire_timer()
151 timer_setup(&this->tl, fsm_expire_timer, 0); in fsm_settimer()
161 del_timer(&this->tl); in fsm_deltimer()
173 timer_setup(&this->tl, fsm_expire_timer, 0); in fsm_addtimer()
176 this->tl.expires = jiffies + (millisec * HZ) / 1000; in fsm_addtimer()
177 add_timer(&this->tl); in fsm_addtimer()
191 del_timer(&this->tl); in fsm_modtimer()
192 timer_setup(&this->tl, fsm_expire_timer, 0); in fsm_modtimer()
195 this->tl.expires = jiffies + (millisec * HZ) / 1000; in fsm_modtimer()
196 add_timer(&this->tl); in fsm_modtimer()
/kernel/linux/linux-5.10/kernel/sched/
Dtopology.c1399 sd_init(struct sched_domain_topology_level *tl, in sd_init() argument
1403 struct sd_data *sdd = &tl->data; in sd_init()
1411 sched_domains_curr_level = tl->numa_level; in sd_init()
1414 sd_weight = cpumask_weight(tl->mask(cpu)); in sd_init()
1416 if (tl->sd_flags) in sd_init()
1417 sd_flags = (*tl->sd_flags)(); in sd_init()
1452 .name = tl->name, in sd_init()
1456 cpumask_and(sched_domain_span(sd), cpu_map, tl->mask(cpu)); in sd_init()
1480 if (sched_domains_numa_distance[tl->numa_level] > node_reclaim_distance) { in sd_init()
1523 #define for_each_sd_topology(tl) \ argument
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/
Di915_request.c341 struct intel_timeline * const tl = i915_request_timeline(rq); in i915_request_retire_upto() local
349 tmp = list_first_entry(&tl->requests, typeof(*tmp), link); in i915_request_retire_upto()
722 static void retire_requests(struct intel_timeline *tl) in retire_requests() argument
726 list_for_each_entry_safe(rq, rn, &tl->requests, link) in retire_requests()
732 request_alloc_slow(struct intel_timeline *tl, in request_alloc_slow() argument
747 if (list_empty(&tl->requests)) in request_alloc_slow()
751 rq = list_first_entry(&tl->requests, typeof(*rq), link); in request_alloc_slow()
760 rq = list_last_entry(&tl->requests, typeof(*rq), link); in request_alloc_slow()
764 retire_requests(tl); in request_alloc_slow()
787 struct intel_timeline *tl = ce->timeline; in __i915_request_create() local
[all …]
/kernel/linux/linux-5.10/arch/x86/crypto/
Dcamellia_glue.c799 u32 dw, tl, tr; in camellia_setup_tail() local
910 tl = (subRL[10] >> 32) ^ (subRL[10] & ~subRL[8]); in camellia_setup_tail()
911 dw = tl & (subRL[8] >> 32); /* FL(kl1) */ in camellia_setup_tail()
913 tt = (tr | ((u64)tl << 32)); in camellia_setup_tail()
919 tl = (subRL[7] >> 32) ^ (subRL[7] & ~subRL[9]); in camellia_setup_tail()
920 dw = tl & (subRL[9] >> 32); /* FLinv(kl2) */ in camellia_setup_tail()
922 tt = (tr | ((u64)tl << 32)); in camellia_setup_tail()
930 tl = (subRL[18] >> 32) ^ (subRL[18] & ~subRL[16]); in camellia_setup_tail()
931 dw = tl & (subRL[16] >> 32); /* FL(kl3) */ in camellia_setup_tail()
933 tt = (tr | ((u64)tl << 32)); in camellia_setup_tail()
[all …]
/kernel/linux/linux-5.10/lib/
Dinflate.c590 struct huft *tl, /* literal/length decoder tables */ in inflate_codes() argument
618 if ((e = (t = tl + ((unsigned)b & ml))->e) > 16) in inflate_codes()
771 struct huft *tl; /* literal/length code table */ in inflate_fixed() local
793 if ((i = huft_build(l, 288, 257, cplens, cplext, &tl, &bl)) != 0) { in inflate_fixed()
804 huft_free(tl); in inflate_fixed()
813 if (inflate_codes(tl, td, bl, bd)) { in inflate_fixed()
820 huft_free(tl); in inflate_fixed()
837 struct huft *tl; /* literal/length code table */ in inflate_dynamic() local
901 if ((i = huft_build(ll, 19, 19, NULL, NULL, &tl, &bl)) != 0) in inflate_dynamic()
904 huft_free(tl); in inflate_dynamic()
[all …]
/kernel/linux/linux-5.10/drivers/net/ethernet/qlogic/qed/
Dqed_vf.h55 struct channel_tlv tl; member
62 struct channel_tlv tl; member
74 struct channel_tlv tl; member
123 struct channel_tlv tl; member
232 struct channel_tlv tl; member
345 struct channel_tlv tl; member
353 struct channel_tlv tl; member
359 struct channel_tlv tl; member
365 struct channel_tlv tl; member
377 struct channel_tlv tl; member
[all …]
/kernel/linux/linux-5.10/arch/s390/kvm/
Dgaccess.c31 unsigned long tl : 2; /* Region- or Segment-Table Length */ member
53 unsigned long tl : 2; /* Region-Second-Table Length */ member
68 unsigned long tl : 2; /* Region-Third-Table Length */ member
82 unsigned long tl : 2; /* Segment-Table Length */ member
636 if (vaddr.rfx01 > asce.tl) in guest_translate()
643 if (vaddr.rsx01 > asce.tl) in guest_translate()
650 if (vaddr.rtx01 > asce.tl) in guest_translate()
657 if (vaddr.sx01 > asce.tl) in guest_translate()
674 if (vaddr.rsx01 < rfte.tf || vaddr.rsx01 > rfte.tl) in guest_translate()
692 if (vaddr.rtx01 < rste.tf || vaddr.rtx01 > rste.tl) in guest_translate()
[all …]
/kernel/linux/linux-5.10/drivers/net/wireless/intel/iwlegacy/
D4965-rs.c235 il4965_rs_tl_rm_old_stats(struct il_traffic_load *tl, u32 curr_time) in il4965_rs_tl_rm_old_stats() argument
240 while (tl->queue_count && tl->time_stamp < oldest_time) { in il4965_rs_tl_rm_old_stats()
241 tl->total -= tl->packet_count[tl->head]; in il4965_rs_tl_rm_old_stats()
242 tl->packet_count[tl->head] = 0; in il4965_rs_tl_rm_old_stats()
243 tl->time_stamp += TID_QUEUE_CELL_SPACING; in il4965_rs_tl_rm_old_stats()
244 tl->queue_count--; in il4965_rs_tl_rm_old_stats()
245 tl->head++; in il4965_rs_tl_rm_old_stats()
246 if (tl->head >= TID_QUEUE_MAX_SIZE) in il4965_rs_tl_rm_old_stats()
247 tl->head = 0; in il4965_rs_tl_rm_old_stats()
261 struct il_traffic_load *tl = NULL; in il4965_rs_tl_add_packet() local
[all …]
/kernel/linux/linux-5.10/arch/sparc/kernel/
Detrap_64.S221 rdpr %tl, %g1
223 wrpr %g0, 1, %tl
233 wrpr %g0, 2, %tl
248 wrpr %g0, 3, %tl
258 wrpr %g0, 4, %tl
271 wrpr %g0, 1, %tl
/kernel/linux/linux-5.10/drivers/net/ethernet/broadcom/bnx2x/
Dbnx2x_vfpf.h99 struct channel_tlv tl; member
105 struct channel_tlv tl; member
117 struct channel_tlv tl; member
213 struct channel_tlv tl; member
219 struct channel_tlv tl; member

1234