/include/linux/ |
D | kvm_types.h | 37 typedef u64 gpa_t; 38 typedef u64 gfn_t; 43 typedef u64 hpa_t; 44 typedef u64 hfn_t; 49 u64 generation; 57 u64 generation; 82 u64 remote_tlb_flush; 83 u64 remote_tlb_flush_requests; 87 u64 halt_successful_poll; 88 u64 halt_attempted_poll; [all …]
|
D | timecounter.h | 14 #define CYCLECOUNTER_MASK(bits) (u64)((bits) < 64 ? ((1ULL<<(bits))-1) : -1) 31 u64 (*read)(const struct cyclecounter *cc); 32 u64 mask; 57 u64 cycle_last; 58 u64 nsec; 59 u64 mask; 60 u64 frac; 70 static inline u64 cyclecounter_cyc2ns(const struct cyclecounter *cc, in cyclecounter_cyc2ns() 71 u64 cycles, u64 mask, u64 *frac) in cyclecounter_cyc2ns() 73 u64 ns = (u64) cycles; in cyclecounter_cyc2ns() [all …]
|
D | firmware-map.h | 17 int firmware_map_add_early(u64 start, u64 end, const char *type); 18 int firmware_map_add_hotplug(u64 start, u64 end, const char *type); 19 int firmware_map_remove(u64 start, u64 end, const char *type); 23 static inline int firmware_map_add_early(u64 start, u64 end, const char *type) in firmware_map_add_early() 28 static inline int firmware_map_add_hotplug(u64 start, u64 end, const char *type) in firmware_map_add_hotplug() 33 static inline int firmware_map_remove(u64 start, u64 end, const char *type) in firmware_map_remove()
|
D | vhost_iotlb.h | 10 u64 start; 11 u64 last; 12 u64 size; 13 u64 addr; 19 u64 __subtree_last; 33 int vhost_iotlb_add_range_ctx(struct vhost_iotlb *iotlb, u64 start, u64 last, 34 u64 addr, unsigned int perm, void *opaque); 35 int vhost_iotlb_add_range(struct vhost_iotlb *iotlb, u64 start, u64 last, 36 u64 addr, unsigned int perm); 37 void vhost_iotlb_del_range(struct vhost_iotlb *iotlb, u64 start, u64 last); [all …]
|
D | math64.h | 26 static inline u64 div_u64_rem(u64 dividend, u32 divisor, u32 *remainder) in div_u64_rem() 54 static inline u64 div64_u64_rem(u64 dividend, u64 divisor, u64 *remainder) in div64_u64_rem() 67 static inline u64 div64_u64(u64 dividend, u64 divisor) in div64_u64() 90 static inline u64 div_u64_rem(u64 dividend, u32 divisor, u32 *remainder) in div_u64_rem() 102 extern u64 div64_u64_rem(u64 dividend, u64 divisor, u64 *remainder); 106 extern u64 div64_u64(u64 dividend, u64 divisor); 125 static inline u64 div_u64(u64 dividend, u32 divisor) in div_u64() 145 u32 iter_div_u64_rem(u64 dividend, u32 divisor, u64 *remainder); 151 static inline u64 mul_u32_u32(u32 a, u32 b) in mul_u32_u32() 153 return (u64)a * b; in mul_u32_u32() [all …]
|
D | task_io_accounting.h | 15 u64 rchar; 17 u64 wchar; 19 u64 syscr; 21 u64 syscw; 23 u64 syscfs; 31 u64 read_bytes; 37 u64 write_bytes; 46 u64 cancelled_write_bytes;
|
D | range.h | 7 u64 start; 8 u64 end; 11 static inline u64 range_len(const struct range *range) in range_len() 17 u64 start, u64 end); 21 u64 start, u64 end); 23 void subtract_range(struct range *range, int az, u64 start, u64 end); 30 static inline resource_size_t cap_resource(u64 val) in cap_resource()
|
D | psi_types.h | 92 u64 state_start; 104 u64 size; 107 u64 start_time; 110 u64 start_value; 113 u64 prev_growth; 121 u64 threshold; 142 u64 last_event_time; 158 u64 avg_total[NR_PSI_STATES - 1]; 159 u64 avg_last_update; 160 u64 avg_next_update; [all …]
|
D | psp-sev.h | 101 u64 tmr_address; /* In */ 114 u64 address; /* In */ 127 u64 pek_cert_address; /* In */ 130 u64 oca_cert_address; /* In */ 141 u64 address; /* In */ 152 u64 address; /* In */ 164 u64 pdh_cert_address; /* In */ 167 u64 cert_chain_address; /* In */ 228 u64 dh_cert_address; /* In */ 231 u64 session_address; /* In */ [all …]
|
D | trace_events.h | 110 u64 ts; 554 extern int synth_event_trace_array(struct trace_event_file *file, u64 *vals, 558 extern int synth_event_add_next_val(u64 val, 560 extern int synth_event_add_val(const char *field_name, u64 val, 728 int perf_event_attach_bpf_prog(struct perf_event *event, struct bpf_prog *prog, u64 bpf_cookie); 737 u64 *probe_offset, u64 *probe_addr); 745 perf_event_attach_bpf_prog(struct perf_event *event, struct bpf_prog *prog, u64 bpf_cookie) in perf_event_attach_bpf_prog() 774 const char **buf, u64 *probe_offset, in bpf_get_perf_event_info() 775 u64 *probe_addr) in bpf_get_perf_event_info() 839 u64 *probe_offset, u64 *probe_addr, [all …]
|
D | cper.h | 332 u64 timestamp; 337 u64 record_id; 339 u64 persistence_information; 361 u64 validation_bits; 369 u64 cpu_version; 371 u64 proc_id; 372 u64 target_addr; 373 u64 requestor_id; 374 u64 responder_id; 375 u64 ip; [all …]
|
D | intel-iommu.h | 33 #define VTD_PAGE_MASK (((u64)-1) << VTD_PAGE_SHIFT) 37 #define VTD_STRIDE_MASK (((u64)-1) << VTD_STRIDE_SHIFT) 215 #define DMA_TLB_GLOBAL_FLUSH (((u64)1) << 60) 216 #define DMA_TLB_DSI_FLUSH (((u64)2) << 60) 217 #define DMA_TLB_PSI_FLUSH (((u64)3) << 60) 220 #define DMA_TLB_READ_DRAIN (((u64)1) << 49) 221 #define DMA_TLB_WRITE_DRAIN (((u64)1) << 48) 222 #define DMA_TLB_DID(id) (((u64)((id) & 0xffff)) << 32) 223 #define DMA_TLB_IVT (((u64)1) << 63) 224 #define DMA_TLB_IH_NONLEAF (((u64)1) << 6) [all …]
|
D | hpet.h | 13 u64 hpet_cap; /* capabilities */ 14 u64 res0; /* reserved */ 15 u64 hpet_config; /* configuration */ 16 u64 res1; /* reserved */ 17 u64 hpet_isr; /* interrupt status reg */ 18 u64 res2[25]; /* reserved */ 20 u64 _hpet_mc64; 24 u64 res3; /* reserved */ 26 u64 hpet_config; /* configuration/cap */ 28 u64 _hpet_hc64; [all …]
|
D | btree-128.h | 22 static inline void *btree_lookup128(struct btree_head128 *head, u64 k1, u64 k2) in btree_lookup128() 24 u64 key[2] = {k1, k2}; in btree_lookup128() 29 u64 *k1, u64 *k2) in btree_get_prev128() 31 u64 key[2] = {*k1, *k2}; in btree_get_prev128() 41 static inline int btree_insert128(struct btree_head128 *head, u64 k1, u64 k2, in btree_insert128() 44 u64 key[2] = {k1, k2}; in btree_insert128() 49 static inline int btree_update128(struct btree_head128 *head, u64 k1, u64 k2, in btree_update128() 52 u64 key[2] = {k1, k2}; in btree_update128() 57 static inline void *btree_remove128(struct btree_head128 *head, u64 k1, u64 k2) in btree_remove128() 59 u64 key[2] = {k1, k2}; in btree_remove128() [all …]
|
D | timekeeper_internal.h | 36 u64 mask; 37 u64 cycle_last; 40 u64 xtime_nsec; 42 u64 base_real; 95 u64 xtime_sec; 105 u64 raw_sec; 109 u64 cycle_interval; 110 u64 xtime_interval; 112 u64 raw_interval; 119 u64 ntp_tick;
|
D | kcov.h | 42 void kcov_remote_start(u64 handle); 44 u64 kcov_common_handle(void); 46 static inline void kcov_remote_start_common(u64 id) in kcov_remote_start_common() 51 static inline void kcov_remote_start_usb(u64 id) in kcov_remote_start_usb() 63 static inline void kcov_remote_start_usb_softirq(u64 id) in kcov_remote_start_usb_softirq() 81 static inline void kcov_remote_start(u64 handle) {} in kcov_remote_start() 83 static inline u64 kcov_common_handle(void) in kcov_common_handle() 87 static inline void kcov_remote_start_common(u64 id) {} in kcov_remote_start_common() 88 static inline void kcov_remote_start_usb(u64 id) {} in kcov_remote_start_usb() 89 static inline void kcov_remote_start_usb_softirq(u64 id) {} in kcov_remote_start_usb_softirq()
|
/include/asm-generic/ |
D | hyperv-tlfs.h | 101 volatile u64 tsc_scale; 179 #define HV_PARTITION_ID_SELF ((u64)-1) 326 u64 as_uint64; 328 u64 enable:1; 329 u64 periodic:1; 330 u64 lazy:1; 331 u64 auto_enable:1; 332 u64 apic_vector:8; 333 u64 direct_mode:1; 334 u64 reserved_z0:3; [all …]
|
/include/linux/ceph/ |
D | striper.h | 11 u64 off, u64 len, 12 u64 *objno, u64 *objoff, u32 *xlen); 16 u64 oe_objno; 17 u64 oe_off; 18 u64 oe_len; 36 int ceph_file_to_extents(struct ceph_file_layout *l, u64 off, u64 len, 42 int ceph_iterate_extents(struct ceph_file_layout *l, u64 off, u64 len, 48 u64 fe_off; 49 u64 fe_len; 52 static inline u64 ceph_file_extents_bytes(struct ceph_file_extent *file_extents, in ceph_file_extents_bytes() [all …]
|
/include/rdma/ |
D | opa_vnic.h | 64 u64 unicast; 65 u64 mcastbcast; 66 u64 untagged; 67 u64 vlan; 68 u64 s_64; 69 u64 s_65_127; 70 u64 s_128_255; 71 u64 s_256_511; 72 u64 s_512_1023; 73 u64 s_1024_1518; [all …]
|
/include/drm/ |
D | drm_mm.h | 159 u64 start; 161 u64 size; 169 u64 __subtree_last; 170 u64 hole_size; 171 u64 subtree_max_hole; 200 u64 *start, u64 *end); 229 u64 size; 230 u64 alignment; 231 u64 remainder_mask; 233 u64 range_start; [all …]
|
/include/kvm/ |
D | arm_pmu.h | 40 u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, u64 select_idx); 41 void kvm_pmu_set_counter_value(struct kvm_vcpu *vcpu, u64 select_idx, u64 val); 42 u64 kvm_pmu_valid_counter_mask(struct kvm_vcpu *vcpu); 43 u64 kvm_pmu_get_pmceid(struct kvm_vcpu *vcpu, bool pmceid1); 47 void kvm_pmu_disable_counter_mask(struct kvm_vcpu *vcpu, u64 val); 48 void kvm_pmu_enable_counter_mask(struct kvm_vcpu *vcpu, u64 val); 53 void kvm_pmu_software_increment(struct kvm_vcpu *vcpu, u64 val); 54 void kvm_pmu_handle_pmcr(struct kvm_vcpu *vcpu, u64 val); 55 void kvm_pmu_set_counter_event_type(struct kvm_vcpu *vcpu, u64 data, 56 u64 select_idx); [all …]
|
/include/linux/sched/ |
D | cputime.h | 22 u64 *utime, u64 *stime); 23 extern u64 task_gtime(struct task_struct *t); 26 u64 *utime, u64 *stime) in task_cputime() 32 static inline u64 task_gtime(struct task_struct *t) in task_gtime() 40 u64 *utimescaled, in task_cputime_scaled() 41 u64 *stimescaled) in task_cputime_scaled() 48 u64 *utimescaled, in task_cputime_scaled() 49 u64 *stimescaled) in task_cputime_scaled() 55 extern void task_cputime_adjusted(struct task_struct *p, u64 *ut, u64 *st); 56 extern void thread_group_cputime_adjusted(struct task_struct *p, u64 *ut, u64 *st); [all …]
|
/include/net/9p/ |
D | 9p.h | 359 u64 path; 392 u64 length; 404 u64 st_result_mask; 409 u64 st_nlink; 410 u64 st_rdev; 411 u64 st_size; 412 u64 st_blksize; 413 u64 st_blocks; 414 u64 st_atime_sec; 415 u64 st_atime_nsec; [all …]
|
/include/trace/events/ |
D | osnoise.h | 11 TP_PROTO(struct task_struct *t, u64 start, u64 duration), 17 __field( u64, start ) 18 __field( u64, duration) 39 TP_PROTO(int vector, u64 start, u64 duration), 44 __field( u64, start ) 45 __field( u64, duration) 65 TP_PROTO(int vector, const char *desc, u64 start, u64 duration), 70 __field( u64, start ) 71 __field( u64, duration) 94 TP_PROTO(u64 start, u64 duration), [all …]
|
D | intel_iommu.h | 21 TP_PROTO(struct intel_iommu *iommu, u64 qw0, u64 qw1, u64 qw2, u64 qw3), 26 __field(u64, qw0) 27 __field(u64, qw1) 28 __field(u64, qw2) 29 __field(u64, qw3) 59 u64 dw0, u64 dw1, u64 dw2, u64 dw3, 65 __field(u64, dw0) 66 __field(u64, dw1) 67 __field(u64, dw2) 68 __field(u64, dw3)
|