Home
last modified time | relevance | path

Searched refs:pages (Results 1 – 25 of 105) sorted by relevance

12345

/drivers/gpu/drm/
Ddrm_scatter.c53 for (i = 0; i < entry->pages; i++) { in drm_sg_cleanup()
62 entry->pages * sizeof(*entry->busaddr), DRM_MEM_PAGES); in drm_sg_cleanup()
64 entry->pages * sizeof(*entry->pagelist), DRM_MEM_PAGES); in drm_sg_cleanup()
77 unsigned long pages, i, j; in drm_sg_alloc() local
92 pages = (request->size + PAGE_SIZE - 1) / PAGE_SIZE; in drm_sg_alloc()
93 DRM_DEBUG("size=%ld pages=%ld\n", request->size, pages); in drm_sg_alloc()
95 entry->pages = pages; in drm_sg_alloc()
96 entry->pagelist = drm_alloc(pages * sizeof(*entry->pagelist), in drm_sg_alloc()
103 memset(entry->pagelist, 0, pages * sizeof(*entry->pagelist)); in drm_sg_alloc()
105 entry->busaddr = drm_alloc(pages * sizeof(*entry->busaddr), in drm_sg_alloc()
[all …]
Dati_pcigart.c60 unsigned long pages; in drm_ati_pcigart_cleanup() local
73 pages = (entry->pages <= max_pages) in drm_ati_pcigart_cleanup()
74 ? entry->pages : max_pages; in drm_ati_pcigart_cleanup()
76 for (i = 0; i < pages; i++) { in drm_ati_pcigart_cleanup()
100 unsigned long pages; in drm_ati_pcigart_init() local
133 pages = (entry->pages <= max_pages) in drm_ati_pcigart_init()
134 ? entry->pages : max_pages; in drm_ati_pcigart_init()
138 for (i = 0; i < pages; i++) { in drm_ati_pcigart_init()
Ddrm_agpsupport.c201 unsigned long pages; in drm_agp_alloc() local
211 pages = (request->size + PAGE_SIZE - 1) / PAGE_SIZE; in drm_agp_alloc()
213 if (!(memory = drm_alloc_agp(dev, pages, type))) { in drm_agp_alloc()
221 entry->pages = pages; in drm_agp_alloc()
371 drm_free_agp(entry->memory, entry->pages); in drm_agp_free()
427 size_t pages, u32 type) in drm_agp_allocate_memory() argument
429 return agp_allocate_memory(bridge, pages, type); in drm_agp_allocate_memory()
466 struct page **pages, in drm_agp_bind_pages() argument
485 mem->memory[i] = phys_to_gart(page_to_phys(pages[i])); in drm_agp_bind_pages()
Ddrm_memory.c99 && (agpmem->bound + (agpmem->pages << PAGE_SHIFT)) >= in agp_remap()
126 DRM_AGP_MEM *drm_alloc_agp(struct drm_device * dev, int pages, u32 type) in drm_alloc_agp() argument
128 return drm_agp_allocate_memory(dev->agp->bridge, pages, type); in drm_alloc_agp()
132 int drm_free_agp(DRM_AGP_MEM * handle, int pages) in drm_free_agp() argument
Ddrm_cache.c51 drm_clflush_pages(struct page *pages[], unsigned long num_pages) in drm_clflush_pages() argument
60 drm_clflush_page(*pages++); in drm_clflush_pages()
/drivers/firewire/
Dfw-iso.c41 buffer->pages = kmalloc(page_count * sizeof(buffer->pages[0]), in fw_iso_buffer_init()
43 if (buffer->pages == NULL) in fw_iso_buffer_init()
47 buffer->pages[i] = alloc_page(GFP_KERNEL | GFP_DMA32 | __GFP_ZERO); in fw_iso_buffer_init()
48 if (buffer->pages[i] == NULL) in fw_iso_buffer_init()
51 address = dma_map_page(card->device, buffer->pages[i], in fw_iso_buffer_init()
54 __free_page(buffer->pages[i]); in fw_iso_buffer_init()
57 set_page_private(buffer->pages[i], address); in fw_iso_buffer_init()
64 address = page_private(buffer->pages[j]); in fw_iso_buffer_init()
67 __free_page(buffer->pages[j]); in fw_iso_buffer_init()
69 kfree(buffer->pages); in fw_iso_buffer_init()
[all …]
/drivers/lguest/x86/
Dcore.c83 static void copy_in_guest_info(struct lg_cpu *cpu, struct lguest_pages *pages) in copy_in_guest_info() argument
89 if (__get_cpu_var(last_cpu) != cpu || cpu->last_pages != pages) { in copy_in_guest_info()
91 cpu->last_pages = pages; in copy_in_guest_info()
97 pages->state.host_cr3 = __pa(current->mm->pgd); in copy_in_guest_info()
100 map_switcher_in_guest(cpu, pages); in copy_in_guest_info()
104 pages->state.guest_tss.sp1 = cpu->esp1; in copy_in_guest_info()
105 pages->state.guest_tss.ss1 = cpu->ss1; in copy_in_guest_info()
109 copy_traps(cpu, pages->state.guest_idt, default_idt_entries); in copy_in_guest_info()
113 copy_gdt(cpu, pages->state.guest_gdt); in copy_in_guest_info()
116 copy_gdt_tls(cpu, pages->state.guest_gdt); in copy_in_guest_info()
[all …]
/drivers/net/mlx4/
Den_resources.c69 struct page **pages; in mlx4_en_map_buffer() local
75 pages = kmalloc(sizeof *pages * buf->nbufs, GFP_KERNEL); in mlx4_en_map_buffer()
76 if (!pages) in mlx4_en_map_buffer()
80 pages[i] = virt_to_page(buf->page_list[i].buf); in mlx4_en_map_buffer()
82 buf->direct.buf = vmap(pages, buf->nbufs, VM_MAP, PAGE_KERNEL); in mlx4_en_map_buffer()
83 kfree(pages); in mlx4_en_map_buffer()
Dalloc.c238 struct page **pages; in mlx4_buf_alloc() local
239 pages = kmalloc(sizeof *pages * buf->nbufs, GFP_KERNEL); in mlx4_buf_alloc()
240 if (!pages) in mlx4_buf_alloc()
243 pages[i] = virt_to_page(buf->page_list[i].buf); in mlx4_buf_alloc()
244 buf->direct.buf = vmap(pages, buf->nbufs, VM_MAP, PAGE_KERNEL); in mlx4_buf_alloc()
245 kfree(pages); in mlx4_buf_alloc()
/drivers/dma/
Diovlock.c52 struct page **pages; in dma_pin_iovec_pages() local
78 pages = (struct page **) &local_list->page_list[nr_iovecs]; in dma_pin_iovec_pages()
93 page_list->pages = pages; in dma_pin_iovec_pages()
94 pages += page_list->nr_pages; in dma_pin_iovec_pages()
105 page_list->pages, in dma_pin_iovec_pages()
133 set_page_dirty_lock(page_list->pages[j]); in dma_unpin_iovec_pages()
134 page_cache_release(page_list->pages[j]); in dma_unpin_iovec_pages()
182 page_list->pages[page_idx], in dma_memcpy_to_iovec()
246 page_list->pages[page_idx], in dma_memcpy_pg_to_iovec()
/drivers/usb/storage/
Dalauda.c614 unsigned int page, unsigned int pages, unsigned char *data) in alauda_read_block_raw() argument
619 PBA_ZONE(pba), 0, PBA_LO(pba) + page, pages, 0, MEDIA_PORT(us) in alauda_read_block_raw()
623 pba, page, pages); in alauda_read_block_raw()
631 data, (MEDIA_INFO(us).pagesize + 64) * pages, NULL); in alauda_read_block_raw()
641 unsigned int page, unsigned int pages, unsigned char *data) in alauda_read_block() argument
646 rc = alauda_read_block_raw(us, pba, page, pages, data); in alauda_read_block()
651 for (i = 0; i < pages; i++) { in alauda_read_block()
694 unsigned int page, unsigned int pages, in alauda_write_lba() argument
761 for (i = page; i < page+pages; i++) { in alauda_write_lba()
837 unsigned int pages; in alauda_read_data() local
[all …]
Dsddr55.c167 unsigned short pages; in sddr55_read_data() local
193 pages = min((unsigned int) sectors << info->smallpageshift, in sddr55_read_data()
195 len = pages << info->pageshift; in sddr55_read_data()
199 pages, pba, lba, page); in sddr55_read_data()
215 command[6] = LSB_of(pages << (1 - info->smallpageshift)); in sddr55_read_data()
262 sectors -= pages >> info->smallpageshift; in sddr55_read_data()
288 unsigned short pages; in sddr55_write_data() local
321 pages = min((unsigned int) sectors << info->smallpageshift, in sddr55_write_data()
323 len = pages << info->pageshift; in sddr55_write_data()
331 pages, pba, lba, page); in sddr55_write_data()
[all …]
Dsddr09.c707 unsigned int page, pages; in sddr09_read_data() local
740 pages = min(sectors, info->blocksize - page); in sddr09_read_data()
741 len = pages << info->pageshift; in sddr09_read_data()
757 pages, lba, page); in sddr09_read_data()
769 pages, pba, lba, page); in sddr09_read_data()
775 pages, info->pageshift, buffer, 0); in sddr09_read_data()
786 sectors -= pages; in sddr09_read_data()
821 unsigned int page, unsigned int pages, in sddr09_write_lba() argument
887 for (i = page; i < page+pages; i++) { in sddr09_write_lba()
932 unsigned int lba, maxlba, page, pages; in sddr09_write_data() local
[all …]
/drivers/media/video/
Dvideobuf-dma-sg.c80 videobuf_pages_to_sg(struct page **pages, int nr_pages, int offset) in videobuf_pages_to_sg() argument
85 if (NULL == pages[0]) in videobuf_pages_to_sg()
92 if (PageHighMem(pages[0])) in videobuf_pages_to_sg()
95 sg_set_page(&sglist[0], pages[0], PAGE_SIZE - offset, offset); in videobuf_pages_to_sg()
97 if (NULL == pages[i]) in videobuf_pages_to_sg()
99 if (PageHighMem(pages[i])) in videobuf_pages_to_sg()
101 sg_set_page(&sglist[i], pages[i], PAGE_SIZE, 0); in videobuf_pages_to_sg()
156 dma->pages = kmalloc(dma->nr_pages * sizeof(struct page*), in videobuf_dma_init_user_locked()
158 if (NULL == dma->pages) in videobuf_dma_init_user_locked()
166 dma->pages, NULL); in videobuf_dma_init_user_locked()
[all …]
Dvideobuf-vmalloc.c157 int pages; in __videobuf_iolock() local
174 pages = PAGE_ALIGN(vb->size); in __videobuf_iolock()
189 mem->vmalloc = vmalloc_user(pages); in __videobuf_iolock()
191 printk(KERN_ERR "vmalloc (%d pages) failed\n", pages); in __videobuf_iolock()
195 mem->vmalloc, pages); in __videobuf_iolock()
256 int retval, pages; in __videobuf_mmap_mapper() local
295 pages = PAGE_ALIGN(vma->vm_end - vma->vm_start); in __videobuf_mmap_mapper()
296 mem->vmalloc = vmalloc_user(pages); in __videobuf_mmap_mapper()
298 printk(KERN_ERR "vmalloc (%d pages) failed\n", pages); in __videobuf_mmap_mapper()
302 mem->vmalloc, pages); in __videobuf_mmap_mapper()
/drivers/md/
Ddm-kcopyd.c36 struct page_list *pages; member
94 unsigned int nr, struct page_list **pages) in kcopyd_get_pages() argument
105 for (*pages = pl = kc->pages; --nr; pl = pl->next) in kcopyd_get_pages()
108 kc->pages = pl->next; in kcopyd_get_pages()
125 cursor->next = kc->pages; in kcopyd_put_pages()
126 kc->pages = pl; in kcopyd_put_pages()
168 drop_pages(kc->pages); in client_free_pages()
169 kc->pages = NULL; in client_free_pages()
203 struct page_list *pages; member
300 kcopyd_put_pages(kc, job->pages); in run_complete_job()
[all …]
Dbitmap.c113 if (page >= bitmap->pages) { in bitmap_checkpage()
116 bmname(bitmap), page, bitmap->pages-1); in bitmap_checkpage()
696 int pages; in bitmap_file_unmap() local
704 pages = bitmap->file_pages; in bitmap_file_unmap()
710 while (pages--) in bitmap_file_unmap()
711 if (map[pages]->index != 0) /* 0 is sb_page, release it below */ in bitmap_file_unmap()
712 free_buffers(map[pages]); in bitmap_file_unmap()
1521 unsigned long k, pages; in bitmap_free() local
1531 pages = bitmap->pages; in bitmap_free()
1536 for (k = 0; k < pages; k++) in bitmap_free()
[all …]
/drivers/lguest/
Dlguest_device.c25 static inline void *lguest_map(unsigned long phys_addr, unsigned long pages) in lguest_map() argument
27 return (__force void *)ioremap_cache(phys_addr, PAGE_SIZE*pages); in lguest_map()
200 void *pages; member
254 lvq->pages = lguest_map((unsigned long)lvq->config.pfn << PAGE_SHIFT, in lg_find_vq()
258 if (!lvq->pages) { in lg_find_vq()
266 vdev, lvq->pages, lg_notify, callback); in lg_find_vq()
293 lguest_unmap(lvq->pages); in lg_find_vq()
310 lguest_unmap(lvq->pages); in lg_del_vq()
Dpage_tables.c706 void map_switcher_in_guest(struct lg_cpu *cpu, struct lguest_pages *pages) in map_switcher_in_guest() argument
728 switcher_pte_page[(unsigned long)pages/PAGE_SIZE%PTRS_PER_PTE] = regs_pte; in map_switcher_in_guest()
746 unsigned int pages) in populate_switcher_pte_page() argument
752 for (i = 0; i < pages; i++) { in populate_switcher_pte_page()
758 i = pages + cpu*2; in populate_switcher_pte_page()
783 __init int init_pagetables(struct page **switcher_page, unsigned int pages) in init_pagetables() argument
793 populate_switcher_pte_page(i, switcher_page, pages); in init_pagetables()
/drivers/mtd/nand/
Dnand_bbt.c220 …res = read_bbt (mtd, buf, td->pages[i], this->chipsize >> this->bbt_erase_shift, bits, offs, td->r… in read_abs_bbt()
226 …res = read_bbt (mtd, buf, td->pages[0], mtd->size >> this->bbt_erase_shift, bits, 0, td->reserved_… in read_abs_bbt()
287 scan_read_raw(mtd, buf, (loff_t)td->pages[0] << this->page_shift, in read_abs_bbts()
291 td->pages[0], td->version[0]); in read_abs_bbts()
296 scan_read_raw(mtd, buf, (loff_t)md->pages[0] << this->page_shift, in read_abs_bbts()
300 md->pages[0], md->version[0]); in read_abs_bbts()
493 td->pages[i] = -1; in search_bbt()
503 td->pages[i] = actblock << blocktopage; in search_bbt()
514 if (td->pages[i] == -1) in search_bbt()
517 printk(KERN_DEBUG "Bad block table found at page %d, version 0x%02X\n", td->pages[i], in search_bbt()
[all …]
/drivers/infiniband/hw/mlx4/
Dmr.c80 u64 *pages; in mlx4_ib_umem_write_mtt() local
87 pages = (u64 *) __get_free_page(GFP_KERNEL); in mlx4_ib_umem_write_mtt()
88 if (!pages) in mlx4_ib_umem_write_mtt()
97 pages[i++] = sg_dma_address(&chunk->page_list[j]) + in mlx4_ib_umem_write_mtt()
105 i, pages); in mlx4_ib_umem_write_mtt()
115 err = mlx4_write_mtt(dev->dev, mtt, n, i, pages); in mlx4_ib_umem_write_mtt()
118 free_page((unsigned long) pages); in mlx4_ib_umem_write_mtt()
/drivers/staging/android/
Dlowmemorykiller.txt4 number of free pages in /sys/module/lowmemorykiller/parameters/minfree. Both
9 with a oom_adj value of 8 or higher when the free memory drops below 4096 pages
11 drops below 1024 pages.
/drivers/net/benet/
Dbe_cmds.c195 static void be_cmd_page_addrs_prepare(struct phys_addr *pages, u32 max_pages, in be_cmd_page_addrs_prepare() argument
202 pages[i].lo = cpu_to_le32(dma & 0xFFFFFFFF); in be_cmd_page_addrs_prepare()
203 pages[i].hi = cpu_to_le32(upper_32_bits(dma)); in be_cmd_page_addrs_prepare()
268 be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem); in be_cmd_eq_create()
395 be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem); in be_cmd_cq_create()
441 be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem); in be_cmd_txq_create()
474 be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem); in be_cmd_rxq_create()
/drivers/block/
Dbrd.c144 struct page *pages[FREE_BATCH]; in brd_free_pages() local
151 (void **)pages, pos, FREE_BATCH); in brd_free_pages()
156 BUG_ON(pages[i]->index < pos); in brd_free_pages()
157 pos = pages[i]->index; in brd_free_pages()
159 BUG_ON(!ret || ret != pages[i]); in brd_free_pages()
160 __free_page(pages[i]); in brd_free_pages()
/drivers/virtio/
Dvirtio_balloon.c47 struct list_head pages; member
114 list_add(&page->lru, &vb->pages); in fill_balloon()
142 page = list_first_entry(&vb->pages, struct page, lru); in leak_balloon()
214 INIT_LIST_HEAD(&vb->pages); in virtballoon_probe()

12345