/arch/x86/um/ |
D | tls_32.c | 97 struct uml_tls_struct* curr = in load_TLS() local 104 if (!curr->present) { in load_TLS() 105 if (!curr->flushed) { in load_TLS() 106 clear_user_desc(&curr->tls); in load_TLS() 107 curr->tls.entry_number = idx; in load_TLS() 109 WARN_ON(!LDT_empty(&curr->tls)); in load_TLS() 114 if (!(flags & O_FORCE) && curr->flushed) in load_TLS() 117 ret = do_set_thread_area(&curr->tls); in load_TLS() 121 curr->flushed = 1; in load_TLS() 137 struct uml_tls_struct* curr = in needs_TLS_update() local [all …]
|
/arch/mips/kernel/ |
D | csrc-r4k.c | 41 unsigned int prev, curr, i; in rdhwr_count_usable() local 50 curr = rdhwr_count(); in rdhwr_count_usable() 52 if (curr != prev) in rdhwr_count_usable() 55 prev = curr; in rdhwr_count_usable()
|
/arch/ia64/kernel/ |
D | unwind.c | 624 memcpy(rs, &sr->curr, sizeof(*rs)); in push() 625 sr->curr.next = rs; in push() 631 struct unw_reg_state *rs = sr->curr.next; in pop() 637 memcpy(&sr->curr, rs, sizeof(*rs)); in pop() 756 reg = sr->curr.reg + unw.save_order[i]; in finish_prologue() 777 regs[0] = sr->curr.reg + UNW_REG_F2; in finish_prologue() 778 regs[1] = sr->curr.reg + UNW_REG_R4; in finish_prologue() 779 regs[2] = sr->curr.reg + UNW_REG_B1; in finish_prologue() 786 spill_next_when(®s[kind - 1], sr->curr.reg + limit[kind - 1], in finish_prologue() 795 alloc_spill_area(&off, 16, sr->curr.reg + UNW_REG_F2, sr->curr.reg + UNW_REG_F31); in finish_prologue() [all …]
|
D | unwind_i.h | 114 struct unw_reg_state curr; /* current state */ member
|
D | mca.c | 500 const struct mca_table_entry *curr; in search_mca_table() local 503 curr = first; in search_mca_table() 504 while (curr <= last) { in search_mca_table() 505 curr_start = (u64) &curr->start_addr + curr->start_addr; in search_mca_table() 506 curr_end = (u64) &curr->end_addr + curr->end_addr; in search_mca_table() 511 curr++; in search_mca_table()
|
/arch/s390/oprofile/ |
D | hwsampler.c | 335 unsigned long *curr; in deallocate_sdbt() local 344 curr = (unsigned long *) sdbt; in deallocate_sdbt() 349 if (!*curr || !sdbt) in deallocate_sdbt() 353 if (is_link_entry(curr)) { in deallocate_sdbt() 354 curr = get_next_sdbt(curr); in deallocate_sdbt() 359 if ((unsigned long) curr == start) in deallocate_sdbt() 362 sdbt = (unsigned long) curr; in deallocate_sdbt() 365 if (*curr) { in deallocate_sdbt() 366 free_page(*curr); in deallocate_sdbt() 367 curr++; in deallocate_sdbt()
|
/arch/x86/kernel/cpu/mtrr/ |
D | generic.c | 82 static int check_type_overlap(u8 *prev, u8 *curr) in check_type_overlap() argument 84 if (*prev == MTRR_TYPE_UNCACHABLE || *curr == MTRR_TYPE_UNCACHABLE) { in check_type_overlap() 86 *curr = MTRR_TYPE_UNCACHABLE; in check_type_overlap() 90 if ((*prev == MTRR_TYPE_WRBACK && *curr == MTRR_TYPE_WRTHROUGH) || in check_type_overlap() 91 (*prev == MTRR_TYPE_WRTHROUGH && *curr == MTRR_TYPE_WRBACK)) { in check_type_overlap() 93 *curr = MTRR_TYPE_WRTHROUGH; in check_type_overlap() 96 if (*prev != *curr) { in check_type_overlap() 98 *curr = MTRR_TYPE_UNCACHABLE; in check_type_overlap()
|
/arch/s390/kernel/ |
D | perf_cpum_sf.c | 118 unsigned long *sdbt, *curr; in free_sampling_buffer() local 124 curr = sdbt; in free_sampling_buffer() 128 if (!*curr || !sdbt) in free_sampling_buffer() 132 if (is_link_entry(curr)) { in free_sampling_buffer() 133 curr = get_next_sdbt(curr); in free_sampling_buffer() 138 if (curr == sfb->sdbt) in free_sampling_buffer() 141 sdbt = curr; in free_sampling_buffer() 144 if (*curr) { in free_sampling_buffer() 145 free_page(*curr); in free_sampling_buffer() 146 curr++; in free_sampling_buffer()
|
/arch/mips/include/asm/ |
D | sgialib.h | 39 extern struct linux_mdesc *prom_getmdesc(struct linux_mdesc *curr);
|
/arch/frv/include/asm/ |
D | processor.h | 44 struct task_struct *curr; /* [GR29] current pointer for this thread */ member
|
/arch/arm/mach-s3c24xx/include/mach/ |
D | dma.h | 204 struct s3c2410_dma_buf *curr; /* current dma buffer */ member
|
/arch/arm/mach-s3c24xx/ |
D | dma.c | 113 chan->curr, chan->next, chan->end); in dmadbg_showchan() 463 if (chan->curr == NULL) { in s3c2410_dma_enqueue() 468 chan->curr = buf; in s3c2410_dma_enqueue() 584 buf = chan->curr; in s3c2410_dma_irq() 634 chan->curr = buf->next; in s3c2410_dma_irq() 899 buf = chan->curr; in s3c2410_dma_flush() 903 chan->curr = chan->next = chan->end = NULL; in s3c2410_dma_flush()
|
/arch/frv/kernel/ |
D | asm-offsets.c | 90 OFFSET(__THREAD_CURR, thread_struct, curr); in foo()
|
D | process.c | 141 p->thread.curr = p; in copy_thread()
|
/arch/powerpc/kernel/ |
D | vio.c | 100 size_t curr; member 165 vio_cmo.curr += size; in vio_cmo_alloc() 166 if (vio_cmo.curr > vio_cmo.high) in vio_cmo_alloc() 167 vio_cmo.high = vio_cmo.curr; in vio_cmo_alloc() 200 vio_cmo.curr -= size; in vio_cmo_dealloc() 1030 viobus_cmo_rd_attr(curr); 1046 vio_cmo.high = vio_cmo.curr; in cmo_high_store()
|
/arch/x86/kernel/cpu/ |
D | common.c | 1401 struct task_struct *curr = current; in cpu_init() local 1403 struct thread_struct *thread = &curr->thread; in cpu_init() 1427 curr->active_mm = &init_mm; in cpu_init() 1428 BUG_ON(curr->mm); in cpu_init() 1429 enter_lazy_tlb(&init_mm, curr); in cpu_init()
|
/arch/parisc/kernel/ |
D | drivers.c | 666 struct parisc_device *curr = to_parisc_device(dev); in match_parisc_device() local 669 return (curr->hw_path == id); in match_parisc_device()
|
/arch/arm/mach-omap2/ |
D | sram243x.S | 170 ldr r3, [r4] @ get curr value
|
D | sram242x.S | 170 ldr r3, [r4] @ get curr value
|
/arch/x86/kvm/ |
D | x86.c | 124 u64 curr; member 187 if (values->host != values->curr) { in kvm_on_user_return() 189 values->curr = values->host; in kvm_on_user_return() 210 smsr->values[slot].curr = value; in shared_msr_update() 238 if (((value ^ smsr->values[slot].curr) & mask) == 0) in kvm_set_shared_msr() 240 smsr->values[slot].curr = value; in kvm_set_shared_msr()
|