Home
last modified time | relevance | path

Searched refs:_addr (Results 1 – 10 of 10) sorted by relevance

/arch/riscv/include/asm/
Dprocessor.h23 typeof(addr) _addr = (addr); \
24 if ((_addr) == 0 || (IS_ENABLED(CONFIG_COMPAT) && is_compat_task())) \
26 else if ((_addr) >= VA_USER_SV57) \
28 else if ((((_addr) >= VA_USER_SV48)) && (VA_BITS >= VA_BITS_SV48)) \
38 typeof(addr) _addr = (addr); \
41 if ((_addr) == 0 || (IS_ENABLED(CONFIG_COMPAT) && is_compat_task())) \
43 else if (((_addr) >= VA_USER_SV57) && (VA_BITS >= VA_BITS_SV57)) \
45 else if ((((_addr) >= VA_USER_SV48)) && (VA_BITS >= VA_BITS_SV48)) \
Dpage.h194 unsigned long _addr = (unsigned long)vaddr; \
195 (unsigned long)(_addr) >= PAGE_OFFSET && pfn_valid(virt_to_pfn(_addr)); \
/arch/x86/kernel/
Didt.c20 #define G(_vector, _addr, _ist, _type, _dpl, _segment) \ argument
27 .addr = _addr, \
32 #define INTG(_vector, _addr) \ argument
33 G(_vector, _addr, DEFAULT_STACK, GATE_INTERRUPT, DPL0, __KERNEL_CS)
36 #define SYSG(_vector, _addr) \ argument
37 G(_vector, _addr, DEFAULT_STACK, GATE_INTERRUPT, DPL3, __KERNEL_CS)
44 #define ISTG(_vector, _addr, _ist) \ argument
45 G(_vector, _addr, _ist + 1, GATE_INTERRUPT, DPL0, __KERNEL_CS)
47 #define ISTG(_vector, _addr, _ist) INTG(_vector, _addr) argument
Dunwind_orc.c387 static bool stack_access_ok(struct unwind_state *state, unsigned long _addr, in stack_access_ok() argument
391 void *addr = (void *)_addr; in stack_access_ok()
/arch/ia64/include/asm/
Dcacheflush.h33 unsigned long _addr = (unsigned long) page_address(page) + ((user_addr) & ~PAGE_MASK); \
34 flush_icache_range(_addr, _addr + (len)); \
/arch/m68k/include/asm/
Draw_io.h83 (void)({u8 __maybe_unused __w, __v = (b); u32 _addr = ((u32) (addr)); \
84 __w = ((*(__force volatile u8 *) ((_addr | 0x10000) + (__v<<1)))); })
86 (void)({u16 __maybe_unused __w, __v = (w); u32 _addr = ((u32) (addr)); \
87 __w = ((*(__force volatile u16 *) ((_addr & 0xFFFF0000UL) + ((__v & 0xFF)<<1)))); \
88 __w = ((*(__force volatile u16 *) ((_addr | 0x10000) + ((__v >> 8)<<1)))); })
90 (void)({u16 __maybe_unused __w, __v = (w); u32 _addr = ((u32) (addr)); \
91 __w = ((*(__force volatile u16 *) ((_addr & 0xFFFF0000UL) + ((__v >> 8)<<1)))); \
92 __w = ((*(__force volatile u16 *) ((_addr | 0x10000) + ((__v & 0xFF)<<1)))); })
/arch/powerpc/include/asm/
Dpage.h241 unsigned long _addr = (unsigned long)vaddr; \
242 _addr >= PAGE_OFFSET && _addr < (unsigned long)high_memory && \
243 pfn_valid(virt_to_pfn((void *)_addr)); \
Dio.h1018 #define setbits32(_addr, _v) out_be32((_addr), in_be32(_addr) | (_v)) argument
1019 #define clrbits32(_addr, _v) out_be32((_addr), in_be32(_addr) & ~(_v)) argument
1021 #define setbits16(_addr, _v) out_be16((_addr), in_be16(_addr) | (_v)) argument
1022 #define clrbits16(_addr, _v) out_be16((_addr), in_be16(_addr) & ~(_v)) argument
1024 #define setbits8(_addr, _v) out_8((_addr), in_8(_addr) | (_v)) argument
1025 #define clrbits8(_addr, _v) out_8((_addr), in_8(_addr) & ~(_v)) argument
/arch/arm/lib/
Duaccess_with_memcpy.c23 pin_page_for_write(const void __user *_addr, pte_t **ptep, spinlock_t **ptlp) in pin_page_for_write() argument
25 unsigned long addr = (unsigned long)_addr; in pin_page_for_write()
/arch/x86/kvm/mmu/
Dmmu.c163 #define for_each_shadow_entry_using_root(_vcpu, _root, _addr, _walker) \ argument
165 (_root), (_addr)); \
169 #define for_each_shadow_entry(_vcpu, _addr, _walker) \ argument
170 for (shadow_walk_init(&(_walker), _vcpu, _addr); \
174 #define for_each_shadow_entry_lockless(_vcpu, _addr, _walker, spte) \ argument
175 for (shadow_walk_init(&(_walker), _vcpu, _addr); \