Searched refs:cur_pages (Results 1 – 3 of 3) sorted by relevance
807 unsigned long cur_pages; in incr_user_locked_vm() local813 cur_pages = atomic_long_read(&pages->source_user->locked_vm); in incr_user_locked_vm()814 new_pages = cur_pages + npages; in incr_user_locked_vm()817 } while (atomic_long_cmpxchg(&pages->source_user->locked_vm, cur_pages, in incr_user_locked_vm()818 new_pages) != cur_pages); in incr_user_locked_vm()1686 struct page **cur_pages; in iopt_pages_fill_xarray() local1689 cur_pages = out_pages + (span.start_used - start_index); in iopt_pages_fill_xarray()1691 span.last_used, cur_pages); in iopt_pages_fill_xarray()1696 cur_pages = out_pages + (span.start_used - start_index); in iopt_pages_fill_xarray()1698 span.last_used, cur_pages); in iopt_pages_fill_xarray()[all …]
345 struct page **cur_pages = NULL; in mmap_batch_fn() local349 cur_pages = &pages[st->index]; in mmap_batch_fn()354 st->domain, cur_pages); in mmap_batch_fn()
908 uint32_t cur_pages; in r100_copy_blit() local931 cur_pages = num_gpu_pages; in r100_copy_blit()932 if (cur_pages > 8191) { in r100_copy_blit()933 cur_pages = 8191; in r100_copy_blit()935 num_gpu_pages -= cur_pages; in r100_copy_blit()959 radeon_ring_write(ring, cur_pages | (stride_pixels << 16)); in r100_copy_blit()