/drivers/gpu/drm/ |
D | drm_scatter.c | 53 for (i = 0; i < entry->pages; i++) { in drm_sg_cleanup() 62 entry->pages * sizeof(*entry->busaddr), DRM_MEM_PAGES); in drm_sg_cleanup() 64 entry->pages * sizeof(*entry->pagelist), DRM_MEM_PAGES); in drm_sg_cleanup() 77 unsigned long pages, i, j; in drm_sg_alloc() local 92 pages = (request->size + PAGE_SIZE - 1) / PAGE_SIZE; in drm_sg_alloc() 93 DRM_DEBUG("size=%ld pages=%ld\n", request->size, pages); in drm_sg_alloc() 95 entry->pages = pages; in drm_sg_alloc() 96 entry->pagelist = drm_alloc(pages * sizeof(*entry->pagelist), in drm_sg_alloc() 103 memset(entry->pagelist, 0, pages * sizeof(*entry->pagelist)); in drm_sg_alloc() 105 entry->busaddr = drm_alloc(pages * sizeof(*entry->busaddr), in drm_sg_alloc() [all …]
|
D | ati_pcigart.c | 60 unsigned long pages; in drm_ati_pcigart_cleanup() local 73 pages = (entry->pages <= max_pages) in drm_ati_pcigart_cleanup() 74 ? entry->pages : max_pages; in drm_ati_pcigart_cleanup() 76 for (i = 0; i < pages; i++) { in drm_ati_pcigart_cleanup() 100 unsigned long pages; in drm_ati_pcigart_init() local 133 pages = (entry->pages <= max_pages) in drm_ati_pcigart_init() 134 ? entry->pages : max_pages; in drm_ati_pcigart_init() 138 for (i = 0; i < pages; i++) { in drm_ati_pcigart_init()
|
D | drm_agpsupport.c | 201 unsigned long pages; in drm_agp_alloc() local 211 pages = (request->size + PAGE_SIZE - 1) / PAGE_SIZE; in drm_agp_alloc() 213 if (!(memory = drm_alloc_agp(dev, pages, type))) { in drm_agp_alloc() 221 entry->pages = pages; in drm_agp_alloc() 371 drm_free_agp(entry->memory, entry->pages); in drm_agp_free() 427 size_t pages, u32 type) in drm_agp_allocate_memory() argument 429 return agp_allocate_memory(bridge, pages, type); in drm_agp_allocate_memory() 466 struct page **pages, in drm_agp_bind_pages() argument 485 mem->memory[i] = phys_to_gart(page_to_phys(pages[i])); in drm_agp_bind_pages()
|
D | drm_memory.c | 99 && (agpmem->bound + (agpmem->pages << PAGE_SHIFT)) >= in agp_remap() 126 DRM_AGP_MEM *drm_alloc_agp(struct drm_device * dev, int pages, u32 type) in drm_alloc_agp() argument 128 return drm_agp_allocate_memory(dev->agp->bridge, pages, type); in drm_alloc_agp() 132 int drm_free_agp(DRM_AGP_MEM * handle, int pages) in drm_free_agp() argument
|
D | drm_cache.c | 51 drm_clflush_pages(struct page *pages[], unsigned long num_pages) in drm_clflush_pages() argument 60 drm_clflush_page(*pages++); in drm_clflush_pages()
|
/drivers/firewire/ |
D | fw-iso.c | 41 buffer->pages = kmalloc(page_count * sizeof(buffer->pages[0]), in fw_iso_buffer_init() 43 if (buffer->pages == NULL) in fw_iso_buffer_init() 47 buffer->pages[i] = alloc_page(GFP_KERNEL | GFP_DMA32 | __GFP_ZERO); in fw_iso_buffer_init() 48 if (buffer->pages[i] == NULL) in fw_iso_buffer_init() 51 address = dma_map_page(card->device, buffer->pages[i], in fw_iso_buffer_init() 54 __free_page(buffer->pages[i]); in fw_iso_buffer_init() 57 set_page_private(buffer->pages[i], address); in fw_iso_buffer_init() 64 address = page_private(buffer->pages[j]); in fw_iso_buffer_init() 67 __free_page(buffer->pages[j]); in fw_iso_buffer_init() 69 kfree(buffer->pages); in fw_iso_buffer_init() [all …]
|
/drivers/lguest/x86/ |
D | core.c | 83 static void copy_in_guest_info(struct lg_cpu *cpu, struct lguest_pages *pages) in copy_in_guest_info() argument 89 if (__get_cpu_var(last_cpu) != cpu || cpu->last_pages != pages) { in copy_in_guest_info() 91 cpu->last_pages = pages; in copy_in_guest_info() 97 pages->state.host_cr3 = __pa(current->mm->pgd); in copy_in_guest_info() 100 map_switcher_in_guest(cpu, pages); in copy_in_guest_info() 104 pages->state.guest_tss.sp1 = cpu->esp1; in copy_in_guest_info() 105 pages->state.guest_tss.ss1 = cpu->ss1; in copy_in_guest_info() 109 copy_traps(cpu, pages->state.guest_idt, default_idt_entries); in copy_in_guest_info() 113 copy_gdt(cpu, pages->state.guest_gdt); in copy_in_guest_info() 116 copy_gdt_tls(cpu, pages->state.guest_gdt); in copy_in_guest_info() [all …]
|
/drivers/net/mlx4/ |
D | en_resources.c | 69 struct page **pages; in mlx4_en_map_buffer() local 75 pages = kmalloc(sizeof *pages * buf->nbufs, GFP_KERNEL); in mlx4_en_map_buffer() 76 if (!pages) in mlx4_en_map_buffer() 80 pages[i] = virt_to_page(buf->page_list[i].buf); in mlx4_en_map_buffer() 82 buf->direct.buf = vmap(pages, buf->nbufs, VM_MAP, PAGE_KERNEL); in mlx4_en_map_buffer() 83 kfree(pages); in mlx4_en_map_buffer()
|
D | alloc.c | 238 struct page **pages; in mlx4_buf_alloc() local 239 pages = kmalloc(sizeof *pages * buf->nbufs, GFP_KERNEL); in mlx4_buf_alloc() 240 if (!pages) in mlx4_buf_alloc() 243 pages[i] = virt_to_page(buf->page_list[i].buf); in mlx4_buf_alloc() 244 buf->direct.buf = vmap(pages, buf->nbufs, VM_MAP, PAGE_KERNEL); in mlx4_buf_alloc() 245 kfree(pages); in mlx4_buf_alloc()
|
/drivers/dma/ |
D | iovlock.c | 52 struct page **pages; in dma_pin_iovec_pages() local 78 pages = (struct page **) &local_list->page_list[nr_iovecs]; in dma_pin_iovec_pages() 93 page_list->pages = pages; in dma_pin_iovec_pages() 94 pages += page_list->nr_pages; in dma_pin_iovec_pages() 105 page_list->pages, in dma_pin_iovec_pages() 133 set_page_dirty_lock(page_list->pages[j]); in dma_unpin_iovec_pages() 134 page_cache_release(page_list->pages[j]); in dma_unpin_iovec_pages() 182 page_list->pages[page_idx], in dma_memcpy_to_iovec() 246 page_list->pages[page_idx], in dma_memcpy_pg_to_iovec()
|
/drivers/usb/storage/ |
D | alauda.c | 614 unsigned int page, unsigned int pages, unsigned char *data) in alauda_read_block_raw() argument 619 PBA_ZONE(pba), 0, PBA_LO(pba) + page, pages, 0, MEDIA_PORT(us) in alauda_read_block_raw() 623 pba, page, pages); in alauda_read_block_raw() 631 data, (MEDIA_INFO(us).pagesize + 64) * pages, NULL); in alauda_read_block_raw() 641 unsigned int page, unsigned int pages, unsigned char *data) in alauda_read_block() argument 646 rc = alauda_read_block_raw(us, pba, page, pages, data); in alauda_read_block() 651 for (i = 0; i < pages; i++) { in alauda_read_block() 694 unsigned int page, unsigned int pages, in alauda_write_lba() argument 761 for (i = page; i < page+pages; i++) { in alauda_write_lba() 837 unsigned int pages; in alauda_read_data() local [all …]
|
D | sddr55.c | 167 unsigned short pages; in sddr55_read_data() local 193 pages = min((unsigned int) sectors << info->smallpageshift, in sddr55_read_data() 195 len = pages << info->pageshift; in sddr55_read_data() 199 pages, pba, lba, page); in sddr55_read_data() 215 command[6] = LSB_of(pages << (1 - info->smallpageshift)); in sddr55_read_data() 262 sectors -= pages >> info->smallpageshift; in sddr55_read_data() 288 unsigned short pages; in sddr55_write_data() local 321 pages = min((unsigned int) sectors << info->smallpageshift, in sddr55_write_data() 323 len = pages << info->pageshift; in sddr55_write_data() 331 pages, pba, lba, page); in sddr55_write_data() [all …]
|
D | sddr09.c | 707 unsigned int page, pages; in sddr09_read_data() local 740 pages = min(sectors, info->blocksize - page); in sddr09_read_data() 741 len = pages << info->pageshift; in sddr09_read_data() 757 pages, lba, page); in sddr09_read_data() 769 pages, pba, lba, page); in sddr09_read_data() 775 pages, info->pageshift, buffer, 0); in sddr09_read_data() 786 sectors -= pages; in sddr09_read_data() 821 unsigned int page, unsigned int pages, in sddr09_write_lba() argument 887 for (i = page; i < page+pages; i++) { in sddr09_write_lba() 932 unsigned int lba, maxlba, page, pages; in sddr09_write_data() local [all …]
|
/drivers/media/video/ |
D | videobuf-dma-sg.c | 80 videobuf_pages_to_sg(struct page **pages, int nr_pages, int offset) in videobuf_pages_to_sg() argument 85 if (NULL == pages[0]) in videobuf_pages_to_sg() 92 if (PageHighMem(pages[0])) in videobuf_pages_to_sg() 95 sg_set_page(&sglist[0], pages[0], PAGE_SIZE - offset, offset); in videobuf_pages_to_sg() 97 if (NULL == pages[i]) in videobuf_pages_to_sg() 99 if (PageHighMem(pages[i])) in videobuf_pages_to_sg() 101 sg_set_page(&sglist[i], pages[i], PAGE_SIZE, 0); in videobuf_pages_to_sg() 156 dma->pages = kmalloc(dma->nr_pages * sizeof(struct page*), in videobuf_dma_init_user_locked() 158 if (NULL == dma->pages) in videobuf_dma_init_user_locked() 166 dma->pages, NULL); in videobuf_dma_init_user_locked() [all …]
|
D | videobuf-vmalloc.c | 157 int pages; in __videobuf_iolock() local 174 pages = PAGE_ALIGN(vb->size); in __videobuf_iolock() 189 mem->vmalloc = vmalloc_user(pages); in __videobuf_iolock() 191 printk(KERN_ERR "vmalloc (%d pages) failed\n", pages); in __videobuf_iolock() 195 mem->vmalloc, pages); in __videobuf_iolock() 256 int retval, pages; in __videobuf_mmap_mapper() local 295 pages = PAGE_ALIGN(vma->vm_end - vma->vm_start); in __videobuf_mmap_mapper() 296 mem->vmalloc = vmalloc_user(pages); in __videobuf_mmap_mapper() 298 printk(KERN_ERR "vmalloc (%d pages) failed\n", pages); in __videobuf_mmap_mapper() 302 mem->vmalloc, pages); in __videobuf_mmap_mapper()
|
/drivers/md/ |
D | dm-kcopyd.c | 36 struct page_list *pages; member 94 unsigned int nr, struct page_list **pages) in kcopyd_get_pages() argument 105 for (*pages = pl = kc->pages; --nr; pl = pl->next) in kcopyd_get_pages() 108 kc->pages = pl->next; in kcopyd_get_pages() 125 cursor->next = kc->pages; in kcopyd_put_pages() 126 kc->pages = pl; in kcopyd_put_pages() 168 drop_pages(kc->pages); in client_free_pages() 169 kc->pages = NULL; in client_free_pages() 203 struct page_list *pages; member 300 kcopyd_put_pages(kc, job->pages); in run_complete_job() [all …]
|
D | bitmap.c | 113 if (page >= bitmap->pages) { in bitmap_checkpage() 116 bmname(bitmap), page, bitmap->pages-1); in bitmap_checkpage() 696 int pages; in bitmap_file_unmap() local 704 pages = bitmap->file_pages; in bitmap_file_unmap() 710 while (pages--) in bitmap_file_unmap() 711 if (map[pages]->index != 0) /* 0 is sb_page, release it below */ in bitmap_file_unmap() 712 free_buffers(map[pages]); in bitmap_file_unmap() 1521 unsigned long k, pages; in bitmap_free() local 1531 pages = bitmap->pages; in bitmap_free() 1536 for (k = 0; k < pages; k++) in bitmap_free() [all …]
|
/drivers/lguest/ |
D | lguest_device.c | 25 static inline void *lguest_map(unsigned long phys_addr, unsigned long pages) in lguest_map() argument 27 return (__force void *)ioremap_cache(phys_addr, PAGE_SIZE*pages); in lguest_map() 200 void *pages; member 254 lvq->pages = lguest_map((unsigned long)lvq->config.pfn << PAGE_SHIFT, in lg_find_vq() 258 if (!lvq->pages) { in lg_find_vq() 266 vdev, lvq->pages, lg_notify, callback); in lg_find_vq() 293 lguest_unmap(lvq->pages); in lg_find_vq() 310 lguest_unmap(lvq->pages); in lg_del_vq()
|
D | page_tables.c | 706 void map_switcher_in_guest(struct lg_cpu *cpu, struct lguest_pages *pages) in map_switcher_in_guest() argument 728 switcher_pte_page[(unsigned long)pages/PAGE_SIZE%PTRS_PER_PTE] = regs_pte; in map_switcher_in_guest() 746 unsigned int pages) in populate_switcher_pte_page() argument 752 for (i = 0; i < pages; i++) { in populate_switcher_pte_page() 758 i = pages + cpu*2; in populate_switcher_pte_page() 783 __init int init_pagetables(struct page **switcher_page, unsigned int pages) in init_pagetables() argument 793 populate_switcher_pte_page(i, switcher_page, pages); in init_pagetables()
|
/drivers/mtd/nand/ |
D | nand_bbt.c | 220 …res = read_bbt (mtd, buf, td->pages[i], this->chipsize >> this->bbt_erase_shift, bits, offs, td->r… in read_abs_bbt() 226 …res = read_bbt (mtd, buf, td->pages[0], mtd->size >> this->bbt_erase_shift, bits, 0, td->reserved_… in read_abs_bbt() 287 scan_read_raw(mtd, buf, (loff_t)td->pages[0] << this->page_shift, in read_abs_bbts() 291 td->pages[0], td->version[0]); in read_abs_bbts() 296 scan_read_raw(mtd, buf, (loff_t)md->pages[0] << this->page_shift, in read_abs_bbts() 300 md->pages[0], md->version[0]); in read_abs_bbts() 493 td->pages[i] = -1; in search_bbt() 503 td->pages[i] = actblock << blocktopage; in search_bbt() 514 if (td->pages[i] == -1) in search_bbt() 517 printk(KERN_DEBUG "Bad block table found at page %d, version 0x%02X\n", td->pages[i], in search_bbt() [all …]
|
/drivers/infiniband/hw/mlx4/ |
D | mr.c | 80 u64 *pages; in mlx4_ib_umem_write_mtt() local 87 pages = (u64 *) __get_free_page(GFP_KERNEL); in mlx4_ib_umem_write_mtt() 88 if (!pages) in mlx4_ib_umem_write_mtt() 97 pages[i++] = sg_dma_address(&chunk->page_list[j]) + in mlx4_ib_umem_write_mtt() 105 i, pages); in mlx4_ib_umem_write_mtt() 115 err = mlx4_write_mtt(dev->dev, mtt, n, i, pages); in mlx4_ib_umem_write_mtt() 118 free_page((unsigned long) pages); in mlx4_ib_umem_write_mtt()
|
/drivers/staging/android/ |
D | lowmemorykiller.txt | 4 number of free pages in /sys/module/lowmemorykiller/parameters/minfree. Both 9 with a oom_adj value of 8 or higher when the free memory drops below 4096 pages 11 drops below 1024 pages.
|
/drivers/net/benet/ |
D | be_cmds.c | 195 static void be_cmd_page_addrs_prepare(struct phys_addr *pages, u32 max_pages, in be_cmd_page_addrs_prepare() argument 202 pages[i].lo = cpu_to_le32(dma & 0xFFFFFFFF); in be_cmd_page_addrs_prepare() 203 pages[i].hi = cpu_to_le32(upper_32_bits(dma)); in be_cmd_page_addrs_prepare() 268 be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem); in be_cmd_eq_create() 395 be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem); in be_cmd_cq_create() 441 be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem); in be_cmd_txq_create() 474 be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem); in be_cmd_rxq_create()
|
/drivers/block/ |
D | brd.c | 144 struct page *pages[FREE_BATCH]; in brd_free_pages() local 151 (void **)pages, pos, FREE_BATCH); in brd_free_pages() 156 BUG_ON(pages[i]->index < pos); in brd_free_pages() 157 pos = pages[i]->index; in brd_free_pages() 159 BUG_ON(!ret || ret != pages[i]); in brd_free_pages() 160 __free_page(pages[i]); in brd_free_pages()
|
/drivers/virtio/ |
D | virtio_balloon.c | 47 struct list_head pages; member 114 list_add(&page->lru, &vb->pages); in fill_balloon() 142 page = list_first_entry(&vb->pages, struct page, lru); in leak_balloon() 214 INIT_LIST_HEAD(&vb->pages); in virtballoon_probe()
|