/arch/s390/include/asm/ |
D | pci_dma.h | 100 static inline void set_pt_pfaa(unsigned long *entry, void *pfaa) in set_pt_pfaa() argument 102 *entry &= ZPCI_PTE_FLAG_MASK; in set_pt_pfaa() 103 *entry |= ((unsigned long) pfaa & ZPCI_PTE_ADDR_MASK); in set_pt_pfaa() 106 static inline void set_rt_sto(unsigned long *entry, void *sto) in set_rt_sto() argument 108 *entry &= ZPCI_RTE_FLAG_MASK; in set_rt_sto() 109 *entry |= ((unsigned long) sto & ZPCI_RTE_ADDR_MASK); in set_rt_sto() 110 *entry |= ZPCI_TABLE_TYPE_RTX; in set_rt_sto() 113 static inline void set_st_pto(unsigned long *entry, void *pto) in set_st_pto() argument 115 *entry &= ZPCI_STE_FLAG_MASK; in set_st_pto() 116 *entry |= ((unsigned long) pto & ZPCI_STE_ADDR_MASK); in set_st_pto() [all …]
|
/arch/sparc/include/asm/ |
D | spitfire.h | 119 static inline unsigned long spitfire_get_dtlb_data(int entry) in spitfire_get_dtlb_data() argument 125 : "r" (entry << 3), "i" (ASI_DTLB_DATA_ACCESS)); in spitfire_get_dtlb_data() 133 static inline unsigned long spitfire_get_dtlb_tag(int entry) in spitfire_get_dtlb_tag() argument 139 : "r" (entry << 3), "i" (ASI_DTLB_TAG_READ)); in spitfire_get_dtlb_tag() 143 static inline void spitfire_put_dtlb_data(int entry, unsigned long data) in spitfire_put_dtlb_data() argument 148 : "r" (data), "r" (entry << 3), in spitfire_put_dtlb_data() 152 static inline unsigned long spitfire_get_itlb_data(int entry) in spitfire_get_itlb_data() argument 158 : "r" (entry << 3), "i" (ASI_ITLB_DATA_ACCESS)); in spitfire_get_itlb_data() 166 static inline unsigned long spitfire_get_itlb_tag(int entry) in spitfire_get_itlb_tag() argument 172 : "r" (entry << 3), "i" (ASI_ITLB_TAG_READ)); in spitfire_get_itlb_tag() [all …]
|
/arch/x86/kvm/ |
D | cpuid.c | 347 struct kvm_cpuid_entry2 entry; in __kvm_cpu_cap_mask() local 352 &entry.eax, &entry.ebx, &entry.ecx, &entry.edx); in __kvm_cpu_cap_mask() 354 kvm_cpu_caps[leaf] &= *__cpuid_entry_get_reg(&entry, cpuid.reg); in __kvm_cpu_cap_mask() 584 struct kvm_cpuid_entry2 *entry = get_next_cpuid(array); in do_host_cpuid() local 586 if (!entry) in do_host_cpuid() 589 entry->function = function; in do_host_cpuid() 590 entry->index = index; in do_host_cpuid() 591 entry->flags = 0; in do_host_cpuid() 593 cpuid_count(entry->function, entry->index, in do_host_cpuid() 594 &entry->eax, &entry->ebx, &entry->ecx, &entry->edx); in do_host_cpuid() [all …]
|
D | reverse_cpuid.h | 115 static __always_inline u32 *__cpuid_entry_get_reg(struct kvm_cpuid_entry2 *entry, in __cpuid_entry_get_reg() argument 120 return &entry->eax; in __cpuid_entry_get_reg() 122 return &entry->ebx; in __cpuid_entry_get_reg() 124 return &entry->ecx; in __cpuid_entry_get_reg() 126 return &entry->edx; in __cpuid_entry_get_reg() 133 static __always_inline u32 *cpuid_entry_get_reg(struct kvm_cpuid_entry2 *entry, in cpuid_entry_get_reg() argument 138 return __cpuid_entry_get_reg(entry, cpuid.reg); in cpuid_entry_get_reg() 141 static __always_inline u32 cpuid_entry_get(struct kvm_cpuid_entry2 *entry, in cpuid_entry_get() argument 144 u32 *reg = cpuid_entry_get_reg(entry, x86_feature); in cpuid_entry_get() 149 static __always_inline bool cpuid_entry_has(struct kvm_cpuid_entry2 *entry, in cpuid_entry_has() argument [all …]
|
/arch/s390/kernel/ |
D | jump_label.c | 18 static void jump_label_make_nop(struct jump_entry *entry, struct insn *insn) in jump_label_make_nop() argument 22 insn->offset = (jump_entry_target(entry) - jump_entry_code(entry)) >> 1; in jump_label_make_nop() 25 static void jump_label_make_branch(struct jump_entry *entry, struct insn *insn) in jump_label_make_branch() argument 29 insn->offset = (jump_entry_target(entry) - jump_entry_code(entry)) >> 1; in jump_label_make_branch() 32 static void jump_label_bug(struct jump_entry *entry, struct insn *expected, in jump_label_bug() argument 35 unsigned char *ipc = (unsigned char *)jump_entry_code(entry); in jump_label_bug() 51 static void __jump_label_transform(struct jump_entry *entry, in __jump_label_transform() argument 55 void *code = (void *)jump_entry_code(entry); in __jump_label_transform() 59 jump_label_make_nop(entry, &old); in __jump_label_transform() 60 jump_label_make_branch(entry, &new); in __jump_label_transform() [all …]
|
/arch/x86/kernel/ |
D | jump_label.c | 20 int arch_jump_entry_size(struct jump_entry *entry) in arch_jump_entry_size() argument 24 insn_decode_kernel(&insn, (void *)jump_entry_code(entry)); in arch_jump_entry_size() 36 __jump_label_patch(struct jump_entry *entry, enum jump_label_type type) in __jump_label_patch() argument 42 addr = (void *)jump_entry_code(entry); in __jump_label_patch() 43 dest = (void *)jump_entry_target(entry); in __jump_label_patch() 45 size = arch_jump_entry_size(entry); in __jump_label_patch() 83 __jump_label_transform(struct jump_entry *entry, in __jump_label_transform() argument 87 const struct jump_label_patch jlp = __jump_label_patch(entry, type); in __jump_label_transform() 101 text_poke_early((void *)jump_entry_code(entry), jlp.code, jlp.size); in __jump_label_transform() 105 text_poke_bp((void *)jump_entry_code(entry), jlp.code, jlp.size, NULL); in __jump_label_transform() [all …]
|
D | e820.c | 83 struct e820_entry *entry = &table->entries[i]; in _e820__mapped_any() local 85 if (type && entry->type != type) in _e820__mapped_any() 87 if (entry->addr >= end || entry->addr + entry->size <= start) in _e820__mapped_any() 118 struct e820_entry *entry = &e820_table->entries[i]; in __e820__mapped_all() local 120 if (type && entry->type != type) in __e820__mapped_all() 124 if (entry->addr >= end || entry->addr + entry->size <= start) in __e820__mapped_all() 131 if (entry->addr <= start) in __e820__mapped_all() 132 start = entry->addr + entry->size; in __e820__mapped_all() 139 return entry; in __e820__mapped_all() 158 struct e820_entry *entry = __e820__mapped_all(start, end, 0); in e820__get_entry_type() local [all …]
|
/arch/riscv/kernel/ |
D | perf_callchain.c | 13 static unsigned long user_backtrace(struct perf_callchain_entry_ctx *entry, in user_backtrace() argument 35 perf_callchain_store(entry, ra); in user_backtrace() 56 void perf_callchain_user(struct perf_callchain_entry_ctx *entry, in perf_callchain_user() argument 67 perf_callchain_store(entry, regs->epc); in perf_callchain_user() 69 fp = user_backtrace(entry, fp, regs->ra); in perf_callchain_user() 70 while (fp && !(fp & 0x3) && entry->nr < entry->max_stack) in perf_callchain_user() 71 fp = user_backtrace(entry, fp, 0); in perf_callchain_user() 74 static bool fill_callchain(void *entry, unsigned long pc) in fill_callchain() argument 76 return perf_callchain_store(entry, pc) == 0; in fill_callchain() 79 void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, in perf_callchain_kernel() argument [all …]
|
/arch/um/kernel/ |
D | irq.c | 96 static bool irq_do_timetravel_handler(struct irq_entry *entry, in irq_do_timetravel_handler() argument 99 struct irq_reg *reg = &entry->reg[t]; in irq_do_timetravel_handler() 110 reg->timetravel_handler(reg->irq, entry->fd, reg->id, ®->event); in irq_do_timetravel_handler() 120 static bool irq_do_timetravel_handler(struct irq_entry *entry, in irq_do_timetravel_handler() argument 127 static void sigio_reg_handler(int idx, struct irq_entry *entry, enum um_irq_type t, in sigio_reg_handler() argument 131 struct irq_reg *reg = &entry->reg[t]; in sigio_reg_handler() 139 if (irq_do_timetravel_handler(entry, t)) in sigio_reg_handler() 228 static bool update_irq_entry(struct irq_entry *entry) in update_irq_entry() argument 234 events |= entry->reg[i].events; in update_irq_entry() 238 os_add_epoll_fd(events, entry->fd, entry); in update_irq_entry() [all …]
|
/arch/x86/boot/compressed/ |
D | idt_64.c | 10 gate_desc entry; in set_idt_entry() local 12 memset(&entry, 0, sizeof(entry)); in set_idt_entry() 14 entry.offset_low = (u16)(address & 0xffff); in set_idt_entry() 15 entry.segment = __KERNEL_CS; in set_idt_entry() 16 entry.bits.type = GATE_TRAP; in set_idt_entry() 17 entry.bits.p = 1; in set_idt_entry() 18 entry.offset_middle = (u16)((address >> 16) & 0xffff); in set_idt_entry() 19 entry.offset_high = (u32)(address >> 32); in set_idt_entry() 21 memcpy(&boot_idt[vector], &entry, sizeof(entry)); in set_idt_entry()
|
/arch/mips/kernel/ |
D | perf_event.c | 26 static void save_raw_perf_callchain(struct perf_callchain_entry_ctx *entry, in save_raw_perf_callchain() argument 35 perf_callchain_store(entry, addr); in save_raw_perf_callchain() 36 if (entry->nr >= entry->max_stack) in save_raw_perf_callchain() 42 void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, in perf_callchain_kernel() argument 55 save_raw_perf_callchain(entry, sp); in perf_callchain_kernel() 59 perf_callchain_store(entry, pc); in perf_callchain_kernel() 60 if (entry->nr >= entry->max_stack) in perf_callchain_kernel() 65 save_raw_perf_callchain(entry, sp); in perf_callchain_kernel()
|
/arch/arm/kernel/ |
D | jump_label.c | 7 static void __arch_jump_label_transform(struct jump_entry *entry, in __arch_jump_label_transform() argument 11 void *addr = (void *)entry->code; in __arch_jump_label_transform() 15 insn = arm_gen_branch(entry->code, entry->target); in __arch_jump_label_transform() 25 void arch_jump_label_transform(struct jump_entry *entry, in arch_jump_label_transform() argument 28 __arch_jump_label_transform(entry, type, false); in arch_jump_label_transform() 31 void arch_jump_label_transform_static(struct jump_entry *entry, in arch_jump_label_transform_static() argument 34 __arch_jump_label_transform(entry, type, true); in arch_jump_label_transform_static()
|
D | perf_callchain.c | 35 struct perf_callchain_entry_ctx *entry) in user_backtrace() argument 50 perf_callchain_store(entry, buftail.lr); in user_backtrace() 63 perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) in perf_callchain_user() argument 73 perf_callchain_store(entry, regs->ARM_pc); in perf_callchain_user() 80 while ((entry->nr < entry->max_stack) && in perf_callchain_user() 82 tail = user_backtrace(tail, entry); in perf_callchain_user() 94 struct perf_callchain_entry_ctx *entry = data; in callchain_trace() local 95 perf_callchain_store(entry, fr->pc); in callchain_trace() 100 perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) in perf_callchain_kernel() argument 111 walk_stackframe(&fr, callchain_trace, entry); in perf_callchain_kernel()
|
/arch/arm64/kernel/ |
D | perf_callchain.c | 24 struct perf_callchain_entry_ctx *entry) in user_backtrace() argument 43 perf_callchain_store(entry, lr); in user_backtrace() 72 struct perf_callchain_entry_ctx *entry) in compat_user_backtrace() argument 88 perf_callchain_store(entry, buftail.lr); in compat_user_backtrace() 102 void perf_callchain_user(struct perf_callchain_entry_ctx *entry, in perf_callchain_user() argument 112 perf_callchain_store(entry, regs->pc); in perf_callchain_user() 120 while (entry->nr < entry->max_stack && in perf_callchain_user() 122 tail = user_backtrace(tail, entry); in perf_callchain_user() 130 while ((entry->nr < entry->max_stack) && in perf_callchain_user() 132 tail = compat_user_backtrace(tail, entry); in perf_callchain_user() [all …]
|
D | jump_label.c | 13 void arch_jump_label_transform(struct jump_entry *entry, in arch_jump_label_transform() argument 16 void *addr = (void *)jump_entry_code(entry); in arch_jump_label_transform() 20 insn = aarch64_insn_gen_branch_imm(jump_entry_code(entry), in arch_jump_label_transform() 21 jump_entry_target(entry), in arch_jump_label_transform() 30 void arch_jump_label_transform_static(struct jump_entry *entry, in arch_jump_label_transform_static() argument
|
/arch/sh/mm/ |
D | tlbex_32.c | 30 pte_t entry; in handle_tlbmiss() local 56 entry = *pte; in handle_tlbmiss() 57 if (unlikely(pte_none(entry) || pte_not_present(entry))) in handle_tlbmiss() 59 if (unlikely(error_code && !pte_write(entry))) in handle_tlbmiss() 63 entry = pte_mkdirty(entry); in handle_tlbmiss() 64 entry = pte_mkyoung(entry); in handle_tlbmiss() 66 set_pte(pte, entry); in handle_tlbmiss()
|
D | tlb-debugfs.c | 45 unsigned int nentries, entry; in tlb_seq_show() local 93 for (entry = 0; entry < nentries; entry++) { in tlb_seq_show() 100 val = __raw_readl(addr1 | (entry << MMU_TLB_ENTRY_SHIFT)); in tlb_seq_show() 105 val = __raw_readl(addr2 | (entry << MMU_TLB_ENTRY_SHIFT)); in tlb_seq_show() 109 val = __raw_readl(data1 | (entry << MMU_TLB_ENTRY_SHIFT)); in tlb_seq_show() 113 val = __raw_readl(data2 | (entry << MMU_TLB_ENTRY_SHIFT)); in tlb_seq_show() 126 entry, vpn, ppn, asid, in tlb_seq_show() 128 (urb <= entry) ? "W" : "-"); in tlb_seq_show()
|
/arch/csky/kernel/ |
D | perf_callchain.c | 36 struct perf_callchain_entry_ctx *entry) in walk_stackframe() argument 39 perf_callchain_store(entry, fr->lr); in walk_stackframe() 47 static unsigned long user_backtrace(struct perf_callchain_entry_ctx *entry, in user_backtrace() argument 67 perf_callchain_store(entry, lr); in user_backtrace() 86 void perf_callchain_user(struct perf_callchain_entry_ctx *entry, in perf_callchain_user() argument 97 perf_callchain_store(entry, regs->pc); in perf_callchain_user() 105 fp = user_backtrace(entry, fp, regs->lr); in perf_callchain_user() 107 while (fp && !(fp & 0x3) && entry->nr < entry->max_stack) in perf_callchain_user() 108 fp = user_backtrace(entry, fp, 0); in perf_callchain_user() 111 void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, in perf_callchain_kernel() argument [all …]
|
/arch/sparc/mm/ |
D | hugetlbpage.c | 134 static pte_t sun4u_hugepage_shift_to_tte(pte_t entry, unsigned int shift) in sun4u_hugepage_shift_to_tte() argument 136 return entry; in sun4u_hugepage_shift_to_tte() 139 static pte_t sun4v_hugepage_shift_to_tte(pte_t entry, unsigned int shift) in sun4v_hugepage_shift_to_tte() argument 143 pte_val(entry) = pte_val(entry) & ~_PAGE_SZALL_4V; in sun4v_hugepage_shift_to_tte() 148 pte_val(entry) |= _PAGE_PUD_HUGE; in sun4v_hugepage_shift_to_tte() 152 pte_val(entry) |= _PAGE_PMD_HUGE; in sun4v_hugepage_shift_to_tte() 156 pte_val(entry) |= _PAGE_PMD_HUGE; in sun4v_hugepage_shift_to_tte() 159 pte_val(entry) |= _PAGE_PMD_HUGE; in sun4v_hugepage_shift_to_tte() 168 pte_val(entry) = pte_val(entry) | hugepage_size; in sun4v_hugepage_shift_to_tte() 169 return entry; in sun4v_hugepage_shift_to_tte() [all …]
|
/arch/mips/mm/ |
D | pgtable-64.c | 19 unsigned long entry; in pgd_init() local 22 entry = (unsigned long)invalid_pud_table; in pgd_init() 24 entry = (unsigned long)invalid_pmd_table; in pgd_init() 26 entry = (unsigned long)invalid_pte_table; in pgd_init() 33 p[0] = entry; in pgd_init() 34 p[1] = entry; in pgd_init() 35 p[2] = entry; in pgd_init() 36 p[3] = entry; in pgd_init() 37 p[4] = entry; in pgd_init() 39 p[-3] = entry; in pgd_init() [all …]
|
/arch/parisc/kernel/ |
D | alternative.c | 25 struct alt_instr *entry; in apply_alternatives() local 45 for (entry = start; entry < end; entry++, index++) { in apply_alternatives() 50 from = (u32 *)((ulong)&entry->orig_offset + entry->orig_offset); in apply_alternatives() 51 len = entry->len; in apply_alternatives() 52 cond = entry->cond; in apply_alternatives() 53 replacement = entry->replacement; in apply_alternatives() 88 source = (u32 *)((ulong)&entry->replacement + entry->replacement); in apply_alternatives()
|
/arch/x86/kernel/cpu/resctrl/ |
D | monitor.c | 132 struct rmid_entry *entry; in __rmid_entry() local 134 entry = &rmid_ptrs[rmid]; in __rmid_entry() 135 WARN_ON(entry->rmid != rmid); in __rmid_entry() 137 return entry; in __rmid_entry() 158 static bool rmid_dirty(struct rmid_entry *entry) in rmid_dirty() argument 160 u64 val = __rmid_read(entry->rmid, QOS_L3_OCCUP_EVENT_ID); in rmid_dirty() 173 struct rmid_entry *entry; in __check_limbo() local 190 entry = __rmid_entry(nrmid); in __check_limbo() 191 if (force_free || !rmid_dirty(entry)) { in __check_limbo() 192 clear_bit(entry->rmid, d->rmid_busy_llc); in __check_limbo() [all …]
|
/arch/m68k/include/asm/ |
D | sun3mmu.h | 88 register unsigned long entry; in sun3_get_segmap() local 95 entry = c; in sun3_get_segmap() 97 return entry; in sun3_get_segmap() 101 static inline void sun3_put_segmap(unsigned long addr, unsigned long entry) in sun3_put_segmap() argument 107 SET_CONTROL_BYTE (AC_SEGMAP | (addr & SUN3_CONTROL_MASK), entry); in sun3_put_segmap() 116 register unsigned long entry; in sun3_get_pte() local 121 GET_CONTROL_WORD (AC_PAGEMAP | (addr & SUN3_CONTROL_MASK), entry); in sun3_get_pte() 124 return entry; in sun3_get_pte() 128 static inline void sun3_put_pte(unsigned long addr, unsigned long entry) in sun3_put_pte() argument 134 SET_CONTROL_WORD (AC_PAGEMAP | (addr & SUN3_CONTROL_MASK), entry); in sun3_put_pte()
|
/arch/powerpc/kvm/ |
D | book3s_64_vio_hv.c | 222 unsigned long entry, unsigned long *hpa, in iommu_tce_xchg_no_kill_rm() argument 227 ret = tbl->it_ops->xchg_no_kill(tbl, entry, hpa, direction, true); in iommu_tce_xchg_no_kill_rm() 231 __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry); in iommu_tce_xchg_no_kill_rm() 244 unsigned long entry, unsigned long pages) in iommu_tce_kill_rm() argument 247 tbl->it_ops->tce_kill(tbl, entry, pages, true); in iommu_tce_kill_rm() 251 struct iommu_table *tbl, unsigned long entry) in kvmppc_rm_clear_tce() argument 255 unsigned long io_entry = entry << (stt->page_shift - tbl->it_page_shift); in kvmppc_rm_clear_tce() 266 struct iommu_table *tbl, unsigned long entry) in kvmppc_rm_tce_iommu_mapped_dec() argument 270 __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry); in kvmppc_rm_tce_iommu_mapped_dec() 288 struct iommu_table *tbl, unsigned long entry) in kvmppc_rm_tce_iommu_do_unmap() argument [all …]
|
/arch/arm/mach-omap2/ |
D | omap-iommu.c | 58 struct pwrdm_link *entry; in _get_pwrdm() local 64 list_for_each_entry(entry, &cache, node) { in _get_pwrdm() 65 if (entry->dev == dev) { in _get_pwrdm() 66 pwrdm = entry->pwrdm; in _get_pwrdm() 101 entry = kmalloc(sizeof(*entry), GFP_KERNEL); in _get_pwrdm() 102 if (entry) { in _get_pwrdm() 103 entry->dev = dev; in _get_pwrdm() 104 entry->pwrdm = pwrdm; in _get_pwrdm() 106 list_add(&entry->node, &cache); in _get_pwrdm()
|