/arch/x86/lib/ |
D | atomic64_386_32.S | 35 #define v %ecx macro 43 #define v %esi macro 51 #define v %esi macro 61 #define v %ecx macro 69 #define v %ecx macro 79 #define v %ecx macro 87 #define v %ecx macro 100 #define v %esi macro 108 #define v %esi macro 120 #define v %esi macro [all …]
|
/arch/ia64/include/asm/ |
D | atomic.h | 24 #define arch_atomic_read(v) READ_ONCE((v)->counter) argument 25 #define arch_atomic64_read(v) READ_ONCE((v)->counter) argument 27 #define arch_atomic_set(v,i) WRITE_ONCE(((v)->counter), (i)) argument 28 #define arch_atomic64_set(v,i) WRITE_ONCE(((v)->counter), (i)) argument 77 #define arch_atomic_add_return(i,v) \ argument 85 #define arch_atomic_sub_return(i,v) \ argument 93 #define arch_atomic_fetch_add(i,v) \ argument 101 #define arch_atomic_fetch_sub(i,v) \ argument 113 #define arch_atomic_and(i,v) (void)ia64_atomic_fetch_and(i,v) argument 114 #define arch_atomic_or(i,v) (void)ia64_atomic_fetch_or(i,v) argument [all …]
|
/arch/x86/include/asm/ |
D | atomic.h | 23 static __always_inline int arch_atomic_read(const atomic_t *v) in arch_atomic_read() 39 static __always_inline void arch_atomic_set(atomic_t *v, int i) in arch_atomic_set() 51 static __always_inline void arch_atomic_add(int i, atomic_t *v) in arch_atomic_add() 65 static __always_inline void arch_atomic_sub(int i, atomic_t *v) in arch_atomic_sub() 81 static __always_inline bool arch_atomic_sub_and_test(int i, atomic_t *v) in arch_atomic_sub_and_test() 93 static __always_inline void arch_atomic_inc(atomic_t *v) in arch_atomic_inc() 106 static __always_inline void arch_atomic_dec(atomic_t *v) in arch_atomic_dec() 121 static __always_inline bool arch_atomic_dec_and_test(atomic_t *v) in arch_atomic_dec_and_test() 135 static __always_inline bool arch_atomic_inc_and_test(atomic_t *v) in arch_atomic_inc_and_test() 150 static __always_inline bool arch_atomic_add_negative(int i, atomic_t *v) in arch_atomic_add_negative() [all …]
|
D | atomic64_64.h | 20 static inline s64 arch_atomic64_read(const atomic64_t *v) in arch_atomic64_read() 32 static inline void arch_atomic64_set(atomic64_t *v, s64 i) in arch_atomic64_set() 44 static __always_inline void arch_atomic64_add(s64 i, atomic64_t *v) in arch_atomic64_add() 58 static inline void arch_atomic64_sub(s64 i, atomic64_t *v) in arch_atomic64_sub() 74 static inline bool arch_atomic64_sub_and_test(s64 i, atomic64_t *v) in arch_atomic64_sub_and_test() 86 static __always_inline void arch_atomic64_inc(atomic64_t *v) in arch_atomic64_inc() 100 static __always_inline void arch_atomic64_dec(atomic64_t *v) in arch_atomic64_dec() 116 static inline bool arch_atomic64_dec_and_test(atomic64_t *v) in arch_atomic64_dec_and_test() 130 static inline bool arch_atomic64_inc_and_test(atomic64_t *v) in arch_atomic64_inc_and_test() 145 static inline bool arch_atomic64_add_negative(s64 i, atomic64_t *v) in arch_atomic64_add_negative() [all …]
|
D | archrandom.h | 20 static inline bool __must_check rdrand_long(unsigned long *v) in rdrand_long() 34 static inline bool __must_check rdrand_int(unsigned int *v) in rdrand_int() 48 static inline bool __must_check rdseed_long(unsigned long *v) in rdseed_long() 57 static inline bool __must_check rdseed_int(unsigned int *v) in rdseed_int() 73 static inline bool __must_check arch_get_random_long(unsigned long *v) in arch_get_random_long() 78 static inline bool __must_check arch_get_random_int(unsigned int *v) in arch_get_random_int() 83 static inline bool __must_check arch_get_random_seed_long(unsigned long *v) in arch_get_random_seed_long() 88 static inline bool __must_check arch_get_random_seed_int(unsigned int *v) in arch_get_random_seed_int()
|
D | atomic64_32.h | 74 static inline s64 arch_atomic64_cmpxchg(atomic64_t *v, s64 o, s64 n) in arch_atomic64_cmpxchg() 88 static inline s64 arch_atomic64_xchg(atomic64_t *v, s64 n) in arch_atomic64_xchg() 107 static inline void arch_atomic64_set(atomic64_t *v, s64 i) in arch_atomic64_set() 122 static inline s64 arch_atomic64_read(const atomic64_t *v) in arch_atomic64_read() 136 static inline s64 arch_atomic64_add_return(s64 i, atomic64_t *v) in arch_atomic64_add_return() 148 static inline s64 arch_atomic64_sub_return(s64 i, atomic64_t *v) in arch_atomic64_sub_return() 157 static inline s64 arch_atomic64_inc_return(atomic64_t *v) in arch_atomic64_inc_return() 166 static inline s64 arch_atomic64_dec_return(atomic64_t *v) in arch_atomic64_dec_return() 182 static inline s64 arch_atomic64_add(s64 i, atomic64_t *v) in arch_atomic64_add() 197 static inline s64 arch_atomic64_sub(s64 i, atomic64_t *v) in arch_atomic64_sub() [all …]
|
/arch/sparc/include/asm/ |
D | atomic_32.h | 33 #define arch_atomic_set_release(v, i) arch_atomic_set((v), (i)) argument 35 #define arch_atomic_read(v) READ_ONCE((v)->counter) argument 37 #define arch_atomic_add(i, v) ((void)arch_atomic_add_return( (int)(i), (v))) argument 38 #define arch_atomic_sub(i, v) ((void)arch_atomic_add_return(-(int)(i), (v))) argument 40 #define arch_atomic_and(i, v) ((void)arch_atomic_fetch_and((i), (v))) argument 41 #define arch_atomic_or(i, v) ((void)arch_atomic_fetch_or((i), (v))) argument 42 #define arch_atomic_xor(i, v) ((void)arch_atomic_fetch_xor((i), (v))) argument 44 #define arch_atomic_sub_return(i, v) (arch_atomic_add_return(-(int)(i), (v))) argument 45 #define arch_atomic_fetch_sub(i, v) (arch_atomic_fetch_add (-(int)(i), (v))) argument
|
D | atomic_64.h | 17 #define arch_atomic_read(v) READ_ONCE((v)->counter) argument 18 #define arch_atomic64_read(v) READ_ONCE((v)->counter) argument 20 #define arch_atomic_set(v, i) WRITE_ONCE(((v)->counter), (i)) argument 21 #define arch_atomic64_set(v, i) WRITE_ONCE(((v)->counter), (i)) argument 52 #define arch_atomic_cmpxchg(v, o, n) (arch_cmpxchg(&((v)->counter), (o), (n))) in ATOMIC_OPS() argument 59 #define arch_atomic64_cmpxchg(v, o, n) \ argument 61 #define arch_atomic64_xchg(v, new) (arch_xchg(&((v)->counter), new)) argument
|
/arch/powerpc/include/asm/ |
D | atomic.h | 26 static __inline__ int arch_atomic_read(const atomic_t *v) in arch_atomic_read() 35 static __inline__ void arch_atomic_set(atomic_t *v, int i) in arch_atomic_set() 136 static __inline__ int arch_atomic_inc_return_relaxed(atomic_t *v) in arch_atomic_inc_return_relaxed() 152 static __inline__ void arch_atomic_dec(atomic_t *v) in arch_atomic_dec() 167 static __inline__ int arch_atomic_dec_return_relaxed(atomic_t *v) in arch_atomic_dec_return_relaxed() 186 #define arch_atomic_cmpxchg(v, o, n) \ argument 188 #define arch_atomic_cmpxchg_relaxed(v, o, n) \ argument 190 #define arch_atomic_cmpxchg_acquire(v, o, n) \ argument 193 #define arch_atomic_xchg(v, new) \ argument 195 #define arch_atomic_xchg_relaxed(v, new) \ argument [all …]
|
D | archrandom.h | 9 static inline bool __must_check arch_get_random_long(unsigned long *v) in arch_get_random_long() 14 static inline bool __must_check arch_get_random_int(unsigned int *v) in arch_get_random_int() 19 static inline bool __must_check arch_get_random_seed_long(unsigned long *v) in arch_get_random_seed_long() 27 static inline bool __must_check arch_get_random_seed_int(unsigned int *v) in arch_get_random_seed_int()
|
/arch/um/include/asm/ |
D | archrandom.h | 10 static inline bool __must_check arch_get_random_long(unsigned long *v) in arch_get_random_long() 15 static inline bool __must_check arch_get_random_int(unsigned int *v) in arch_get_random_int() 20 static inline bool __must_check arch_get_random_seed_long(unsigned long *v) in arch_get_random_seed_long() 25 static inline bool __must_check arch_get_random_seed_int(unsigned int *v) in arch_get_random_seed_int()
|
/arch/s390/include/asm/ |
D | archrandom.h | 24 static inline bool __must_check arch_get_random_long(unsigned long *v) in arch_get_random_long() 29 static inline bool __must_check arch_get_random_int(unsigned int *v) in arch_get_random_int() 34 static inline bool __must_check arch_get_random_seed_long(unsigned long *v) in arch_get_random_seed_long() 45 static inline bool __must_check arch_get_random_seed_int(unsigned int *v) in arch_get_random_seed_int()
|
D | atomic.h | 18 static inline int arch_atomic_read(const atomic_t *v) in arch_atomic_read() 24 static inline void arch_atomic_set(atomic_t *v, int i) in arch_atomic_set() 30 static inline int arch_atomic_add_return(int i, atomic_t *v) in arch_atomic_add_return() 36 static inline int arch_atomic_fetch_add(int i, atomic_t *v) in arch_atomic_fetch_add() 42 static inline void arch_atomic_add(int i, atomic_t *v) in arch_atomic_add() 75 #define arch_atomic_xchg(v, new) (arch_xchg(&((v)->counter), new)) in ATOMIC_OPS() argument 85 static inline s64 arch_atomic64_read(const atomic64_t *v) in arch_atomic64_read() 91 static inline void arch_atomic64_set(atomic64_t *v, s64 i) in arch_atomic64_set() 97 static inline s64 arch_atomic64_add_return(s64 i, atomic64_t *v) in arch_atomic64_add_return() 103 static inline s64 arch_atomic64_fetch_add(s64 i, atomic64_t *v) in arch_atomic64_fetch_add() [all …]
|
/arch/arc/include/asm/ |
D | atomic.h | 17 #define arch_atomic_read(v) READ_ONCE((v)->counter) argument 25 #define arch_atomic_cmpxchg(v, o, n) \ argument 31 #define arch_atomic_cmpxchg_relaxed(v, o, n) \ argument 37 #define arch_atomic_xchg(v, n) \ argument 43 #define arch_atomic_xchg_relaxed(v, n) \ argument
|
D | setup.h | 30 #define IS_AVAIL1(v, s) ((v) ? s : "") argument 31 #define IS_DISABLED_RUN(v) ((v) ? "" : "(disabled) ") argument 32 #define IS_USED_RUN(v) ((v) ? "" : "(not used) ") argument 34 #define IS_AVAIL2(v, s, cfg) IS_AVAIL1(v, s), IS_AVAIL1(v, IS_USED_CFG(cfg)) argument 35 #define IS_AVAIL3(v, v2, s) IS_AVAIL1(v, s), IS_AVAIL1(v, IS_DISABLED_RUN(v2)) argument
|
/arch/x86/kvm/ |
D | irq.c | 35 static int pending_userspace_extint(struct kvm_vcpu *v) in pending_userspace_extint() 44 int kvm_cpu_has_extint(struct kvm_vcpu *v) in kvm_cpu_has_extint() 78 int kvm_cpu_has_injectable_intr(struct kvm_vcpu *v) in kvm_cpu_has_injectable_intr() 94 int kvm_cpu_has_interrupt(struct kvm_vcpu *v) in kvm_cpu_has_interrupt() 107 static int kvm_cpu_get_extint(struct kvm_vcpu *v) in kvm_cpu_get_extint() 132 int kvm_cpu_get_interrupt(struct kvm_vcpu *v) in kvm_cpu_get_interrupt()
|
/arch/mips/include/asm/mach-bcm63xx/ |
D | bcm63xx_io.h | 58 #define bcm_writeb(v, a) (*(volatile unsigned char *) BCM_REGS_VA((a)) = (v)) argument 59 #define bcm_writew(v, a) (*(volatile unsigned short *) BCM_REGS_VA((a)) = (v)) argument 60 #define bcm_writel(v, a) (*(volatile unsigned int *) BCM_REGS_VA((a)) = (v)) argument 61 #define bcm_writeq(v, a) (*(volatile u64 *) BCM_REGS_VA((a)) = (v)) argument 69 #define bcm_rset_writeb(s, v, o) bcm_writeb((v), \ argument 71 #define bcm_rset_writew(s, v, o) bcm_writew((v), \ argument 73 #define bcm_rset_writel(s, v, o) bcm_writel((v), \ argument 80 #define bcm_perf_writel(v, o) bcm_rset_writel(RSET_PERF, (v), (o)) argument 82 #define bcm_timer_writel(v, o) bcm_rset_writel(RSET_TIMER, (v), (o)) argument 84 #define bcm_wdt_writel(v, o) bcm_rset_writel(RSET_WDT, (v), (o)) argument [all …]
|
/arch/m68k/include/asm/ |
D | atomic.h | 19 #define arch_atomic_read(v) READ_ONCE((v)->counter) argument 20 #define arch_atomic_set(v, i) WRITE_ONCE(((v)->counter), (i)) argument 123 static inline void arch_atomic_inc(atomic_t *v) in arch_atomic_inc() 129 static inline void arch_atomic_dec(atomic_t *v) in arch_atomic_dec() 135 static inline int arch_atomic_dec_and_test(atomic_t *v) in arch_atomic_dec_and_test() 143 static inline int arch_atomic_dec_and_test_lt(atomic_t *v) in arch_atomic_dec_and_test_lt() 153 static inline int arch_atomic_inc_and_test(atomic_t *v) in arch_atomic_inc_and_test() 163 #define arch_atomic_cmpxchg(v, o, n) ((int)arch_cmpxchg(&((v)->counter), (o), (n))) argument 164 #define arch_atomic_xchg(v, new) (arch_xchg(&((v)->counter), new)) argument 168 static inline int arch_atomic_cmpxchg(atomic_t *v, int old, int new) in arch_atomic_cmpxchg() [all …]
|
/arch/arm64/include/asm/ |
D | archrandom.h | 27 static inline bool __arm64_rndr(unsigned long *v) in __arm64_rndr() 45 static inline bool __must_check arch_get_random_long(unsigned long *v) in arch_get_random_long() 50 static inline bool __must_check arch_get_random_int(unsigned int *v) in arch_get_random_int() 55 static inline bool __must_check arch_get_random_seed_long(unsigned long *v) in arch_get_random_seed_long() 85 static inline bool __must_check arch_get_random_seed_int(unsigned int *v) in arch_get_random_seed_int() 116 arch_get_random_seed_long_early(unsigned long *v) in arch_get_random_seed_long_early()
|
/arch/csky/include/asm/ |
D | io.h | 26 #define writeb(v,c) ({ wmb(); writeb_relaxed((v),(c)); }) argument 27 #define writew(v,c) ({ wmb(); writew_relaxed((v),(c)); }) argument 28 #define writel(v,c) ({ wmb(); writel_relaxed((v),(c)); }) argument 30 #define writeb(v,c) ({ wmb(); writeb_relaxed((v),(c)); mb(); }) argument 31 #define writew(v,c) ({ wmb(); writew_relaxed((v),(c)); mb(); }) argument 32 #define writel(v,c) ({ wmb(); writel_relaxed((v),(c)); mb(); }) argument
|
/arch/sh/include/asm/ |
D | atomic.h | 22 #define arch_atomic_read(v) READ_ONCE((v)->counter) argument 23 #define arch_atomic_set(v,i) WRITE_ONCE((v)->counter, (i)) argument 33 #define arch_atomic_xchg(v, new) (arch_xchg(&((v)->counter), new)) argument 34 #define arch_atomic_cmpxchg(v, o, n) (arch_cmpxchg(&((v)->counter), (o), (n))) argument
|
/arch/riscv/kernel/ |
D | module.c | 29 static int apply_r_riscv_32_rela(struct module *me, u32 *location, Elf_Addr v) in apply_r_riscv_32_rela() 40 static int apply_r_riscv_64_rela(struct module *me, u32 *location, Elf_Addr v) in apply_r_riscv_64_rela() 47 Elf_Addr v) in apply_r_riscv_branch_rela() 60 Elf_Addr v) in apply_r_riscv_jal_rela() 73 Elf_Addr v) in apply_r_riscv_rcv_branch_rela() 88 Elf_Addr v) in apply_r_riscv_rvc_jump_rela() 106 Elf_Addr v) in apply_r_riscv_pcrel_hi20_rela() 124 Elf_Addr v) in apply_r_riscv_pcrel_lo12_i_rela() 135 Elf_Addr v) in apply_r_riscv_pcrel_lo12_s_rela() 149 Elf_Addr v) in apply_r_riscv_hi20_rela() [all …]
|
/arch/x86/kernel/ |
D | bootflag.c | 23 static int __init parity(u8 v) in parity() 36 static void __init sbf_write(u8 v) in sbf_write() 57 u8 v; in sbf_read() local 69 static int __init sbf_value_valid(u8 v) in sbf_value_valid() 81 u8 v; in sbf_init() local
|
/arch/x86/mm/ |
D | debug_pagetables.c | 8 static int ptdump_show(struct seq_file *m, void *v) in ptdump_show() 16 static int ptdump_curknl_show(struct seq_file *m, void *v) in ptdump_curknl_show() 26 static int ptdump_curusr_show(struct seq_file *m, void *v) in ptdump_curusr_show() 37 static int ptdump_efi_show(struct seq_file *m, void *v) in ptdump_efi_show()
|
/arch/riscv/include/asm/ |
D | mmio.h | 91 #define writeb_cpu(v, c) ((void)__raw_writeb((v), (c))) argument 92 #define writew_cpu(v, c) ((void)__raw_writew((__force u16)cpu_to_le16(v), (c))) argument 93 #define writel_cpu(v, c) ((void)__raw_writel((__force u32)cpu_to_le32(v), (c))) argument 97 #define writeq_cpu(v, c) ((void)__raw_writeq((__force u64)cpu_to_le64(v), (c))) argument 118 #define writeb_relaxed(v, c) ({ __io_rbw(); writeb_cpu((v), (c)); __io_raw(); }) argument 119 #define writew_relaxed(v, c) ({ __io_rbw(); writew_cpu((v), (c)); __io_raw(); }) argument 120 #define writel_relaxed(v, c) ({ __io_rbw(); writel_cpu((v), (c)); __io_raw(); }) argument 124 #define writeq_relaxed(v, c) ({ __io_rbw(); writeq_cpu((v), (c)); __io_raw(); }) argument 134 #define __io_ar(v) ({ __asm__ __volatile__ ("fence i,ir" : : : "memory"); }) argument 142 #define writeb(v, c) ({ __io_bw(); writeb_cpu((v), (c)); __io_aw(); }) argument [all …]
|