/drivers/media/video/ivtv/ |
D | ivtv-udma.c | 33 dma_page->page_count = dma_page->last - dma_page->first + 1; in ivtv_udma_get_page_info() 34 if (dma_page->page_count == 1) dma_page->tail -= dma_page->offset; in ivtv_udma_get_page_info() 48 for (i = 0; i < dma_page->page_count; i++) { in ivtv_udma_fill_sg_list() 49 unsigned int len = (i == dma_page->page_count - 1) ? in ivtv_udma_fill_sg_list() 112 if (dma->SG_length || dma->page_count) { in ivtv_udma_setup() 114 dma->SG_length, dma->page_count); in ivtv_udma_setup() 120 if (user_dma.page_count <= 0) { in ivtv_udma_setup() 122 user_dma.page_count, size_in_bytes, user_dma.offset); in ivtv_udma_setup() 129 user_dma.uaddr, user_dma.page_count, 0, 1, dma->map, NULL); in ivtv_udma_setup() 132 if (user_dma.page_count != err) { in ivtv_udma_setup() [all …]
|
D | ivtv-yuv.c | 67 if (dma->SG_length || dma->page_count) { in ivtv_yuv_prep_user_dma() 70 dma->SG_length, dma->page_count); in ivtv_yuv_prep_user_dma() 79 …y_pages = get_user_pages(current, current->mm, y_dma.uaddr, y_dma.page_count, 0, 1, &dma->map[0], … in ivtv_yuv_prep_user_dma() 80 …uv_pages = get_user_pages(current, current->mm, uv_dma.uaddr, uv_dma.page_count, 0, 1, &dma->map[y… in ivtv_yuv_prep_user_dma() 83 dma->page_count = y_dma.page_count + uv_dma.page_count; in ivtv_yuv_prep_user_dma() 85 if (y_pages + uv_pages != dma->page_count) { in ivtv_yuv_prep_user_dma() 88 y_pages + uv_pages, dma->page_count); in ivtv_yuv_prep_user_dma() 90 for (i = 0; i < dma->page_count; i++) { in ivtv_yuv_prep_user_dma() 93 dma->page_count = 0; in ivtv_yuv_prep_user_dma() 100 for (i = 0; i < dma->page_count; i++) { in ivtv_yuv_prep_user_dma() [all …]
|
/drivers/firewire/ |
D | fw-iso.c | 33 int page_count, enum dma_data_direction direction) in fw_iso_buffer_init() argument 38 buffer->page_count = page_count; in fw_iso_buffer_init() 41 buffer->pages = kmalloc(page_count * sizeof(buffer->pages[0]), in fw_iso_buffer_init() 46 for (i = 0; i < buffer->page_count; i++) { in fw_iso_buffer_init() 81 for (i = 0; i < buffer->page_count; i++) { in fw_iso_buffer_map() 97 for (i = 0; i < buffer->page_count; i++) { in fw_iso_buffer_destroy()
|
D | fw-transaction.h | 334 int page_count; member 350 int page_count,
|
/drivers/char/agp/ |
D | generic.c | 203 if (curr->page_count != 0) { in agp_free_memory() 208 for (i = 0; i < curr->page_count; i++) { in agp_free_memory() 215 for (i = 0; i < curr->page_count; i++) { in agp_free_memory() 242 size_t page_count, u32 type) in agp_allocate_memory() argument 251 if ((atomic_read(&bridge->current_memory_agp) + page_count) > bridge->max_memory_agp) in agp_allocate_memory() 255 new = agp_generic_alloc_user(page_count, type); in agp_allocate_memory() 262 new = bridge->driver->alloc_by_type(page_count, type); in agp_allocate_memory() 268 scratch_pages = (page_count + ENTRIES_PER_PAGE - 1) / ENTRIES_PER_PAGE; in agp_allocate_memory() 276 if (bridge->driver->agp_alloc_pages(bridge, new, page_count)) { in agp_allocate_memory() 284 for (i = 0; i < page_count; i++) { in agp_allocate_memory() [all …]
|
D | uninorth-agp.c | 156 if ((pg_start + mem->page_count) > num_entries) in uninorth_insert_memory() 161 while (j < (pg_start + mem->page_count)) { in uninorth_insert_memory() 167 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { in uninorth_insert_memory() 176 (unsigned long)&agp_bridge->gatt_table[pg_start + mem->page_count]); in uninorth_insert_memory() 194 if ((pg_start + mem->page_count) > num_entries) in u3_insert_memory() 198 for (i = 0; i < mem->page_count; ++i) { in u3_insert_memory() 207 for (i = 0; i < mem->page_count; i++) { in u3_insert_memory() 229 for (i = 0; i < mem->page_count; ++i) in u3_remove_memory()
|
D | i460-agp.c | 307 if ((io_pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count) > num_entries) { in i460_insert_memory_small_io_page() 313 while (j < (io_pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count)) { in i460_insert_memory_small_io_page() 323 for (i = 0, j = io_pg_start; i < mem->page_count; i++) { in i460_insert_memory_small_io_page() 343 for (i = pg_start; i < (pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count); i++) in i460_remove_memory_small_io_page() 413 end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; in i460_insert_memory_large_io_page() 415 end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE; in i460_insert_memory_large_io_page() 471 end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; in i460_remove_memory_large_io_page() 473 end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE; in i460_remove_memory_large_io_page()
|
D | sgi-agp.c | 174 if ((pg_start + mem->page_count) > num_entries) in sgi_tioca_insert_memory() 179 while (j < (pg_start + mem->page_count)) { in sgi_tioca_insert_memory() 190 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { in sgi_tioca_insert_memory() 217 for (i = pg_start; i < (mem->page_count + pg_start); i++) { in sgi_tioca_remove_memory()
|
D | nvidia-agp.c | 210 if (mem->page_count == 0) in nvidia_insert_memory() 213 if ((pg_start + mem->page_count) > in nvidia_insert_memory() 217 for (j = pg_start; j < (pg_start + mem->page_count); j++) { in nvidia_insert_memory() 226 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { in nvidia_insert_memory() 250 if (mem->page_count == 0) in nvidia_remove_memory() 253 for (i = pg_start; i < (mem->page_count + pg_start); i++) in nvidia_remove_memory()
|
D | intel-agp.c | 296 if (mem->page_count == 0) in intel_i810_insert_entries() 302 if ((pg_start + mem->page_count) > num_entries) in intel_i810_insert_entries() 306 for (j = pg_start; j < (pg_start + mem->page_count); j++) { in intel_i810_insert_entries() 322 for (i = pg_start; i < (pg_start + mem->page_count); i++) { in intel_i810_insert_entries() 332 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { in intel_i810_insert_entries() 357 if (mem->page_count == 0) in intel_i810_remove_entries() 360 for (i = pg_start; i < (mem->page_count + pg_start); i++) { in intel_i810_remove_entries() 404 new->page_count = pg_count; in alloc_agpphysmem_i8xx() 424 new->page_count = pg_count; in intel_i810_alloc_by_type() 438 if (curr->page_count == 4) in intel_i810_free_by_type() [all …]
|
D | efficeon-agp.c | 239 int i, count = mem->page_count, num_entries; in efficeon_insert_memory() 247 if ((pg_start + mem->page_count) > num_entries) in efficeon_insert_memory() 288 int i, count = mem->page_count, num_entries; in efficeon_remove_memory() 294 if ((pg_start + mem->page_count) > num_entries) in efficeon_remove_memory()
|
D | agp.h | 279 struct agp_memory *agp_generic_alloc_by_type(size_t page_count, int type); 283 struct agp_memory *memory, size_t page_count); 300 struct agp_memory *agp_generic_alloc_user(size_t page_count, int type);
|
D | ati-agp.c | 278 if ((pg_start + mem->page_count) > num_entries) in ati_insert_memory() 282 while (j < (pg_start + mem->page_count)) { in ati_insert_memory() 296 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { in ati_insert_memory() 317 for (i = pg_start; i < (mem->page_count + pg_start); i++) { in ati_remove_memory()
|
D | parisc-agp.c | 132 io_pg_count = info->io_pages_per_kpage * mem->page_count; in parisc_agp_insert_memory() 149 for (i = 0, j = io_pg_start; i < mem->page_count; i++) { in parisc_agp_insert_memory() 178 io_pg_count = info->io_pages_per_kpage * mem->page_count; in parisc_agp_remove_memory()
|
D | ali-agp.c | 128 int i, page_count; in m1541_cache_flush() local 133 page_count = 1 << A_SIZE_32(agp_bridge->current_size)->page_order; in m1541_cache_flush() 134 for (i = 0; i < PAGE_SIZE * page_count; i += PAGE_SIZE) { in m1541_cache_flush()
|
D | sworks-agp.c | 331 if ((pg_start + mem->page_count) > num_entries) { in serverworks_insert_memory() 336 while (j < (pg_start + mem->page_count)) { in serverworks_insert_memory() 349 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { in serverworks_insert_memory() 372 for (i = pg_start; i < (mem->page_count + pg_start); i++) { in serverworks_remove_memory()
|
D | amd-k7-agp.c | 307 if ((pg_start + mem->page_count) > num_entries) in amd_insert_memory() 311 while (j < (pg_start + mem->page_count)) { in amd_insert_memory() 324 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { in amd_insert_memory() 344 for (i = pg_start; i < (mem->page_count + pg_start); i++) { in amd_remove_memory()
|
D | hp-agp.c | 343 io_pg_count = hp->io_pages_per_kpage * mem->page_count; in hp_zx1_insert_memory() 361 for (i = 0, j = io_pg_start; i < mem->page_count; i++) { in hp_zx1_insert_memory() 389 io_pg_count = hp->io_pages_per_kpage * mem->page_count; in hp_zx1_remove_memory()
|
D | amd64-agp.c | 63 if (((unsigned long)pg_start + mem->page_count) > num_entries) in amd64_insert_memory() 69 while (j < (pg_start + mem->page_count)) { in amd64_insert_memory() 80 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { in amd64_insert_memory()
|
/drivers/gpu/drm/ |
D | drm_bufs.c | 710 dma->page_count += byte_count >> PAGE_SHIFT; in drm_addbufs_agp() 744 int page_count; in drm_addbufs_pci() local 819 temp_pagelist = drm_alloc((dma->page_count + (count << page_order)) in drm_addbufs_pci() 831 dma->pagelist, dma->page_count * sizeof(*dma->pagelist)); in drm_addbufs_pci() 833 dma->page_count + (count << page_order)); in drm_addbufs_pci() 838 page_count = 0; in drm_addbufs_pci() 850 (dma->page_count + (count << page_order)) in drm_addbufs_pci() 859 dma->page_count + page_count, in drm_addbufs_pci() 861 temp_pagelist[dma->page_count + page_count++] in drm_addbufs_pci() 890 (dma->page_count + in drm_addbufs_pci() [all …]
|
/drivers/gpu/drm/i915/ |
D | i915_gem.c | 822 int page_count = obj->size / PAGE_SIZE; in i915_gem_object_free_page_list() local 829 for (i = 0; i < page_count; i++) in i915_gem_object_free_page_list() 839 page_count * sizeof(struct page *), in i915_gem_object_free_page_list() 1415 int page_count, i; in i915_gem_object_get_page_list() local 1427 page_count = obj->size / PAGE_SIZE; in i915_gem_object_get_page_list() 1429 obj_priv->page_list = drm_calloc(page_count, sizeof(struct page *), in i915_gem_object_get_page_list() 1438 for (i = 0; i < page_count; i++) { in i915_gem_object_get_page_list() 1720 int page_count, ret; in i915_gem_object_bind_to_gtt() local 1776 page_count = obj->size / PAGE_SIZE; in i915_gem_object_bind_to_gtt() 1782 page_count, in i915_gem_object_bind_to_gtt() [all …]
|
/drivers/block/ |
D | ps3vram.c | 59 unsigned int page_count; member 360 for (i = 0; i < cache->page_count; i++) { in ps3vram_cache_flush() 380 for (i = 0; i < cache->page_count; i++) { in ps3vram_cache_match() 391 i = (jiffies + (counter++)) % cache->page_count; in ps3vram_cache_match() 405 priv->cache.page_count = CACHE_PAGE_COUNT; in ps3vram_cache_init()
|
/drivers/staging/go7007/ |
D | go7007-v4l2.c | 57 if (gobuf->page_count > 0) { in deactivate_buffer() 58 for (i = 0; i < gobuf->page_count; ++i) in deactivate_buffer() 60 gobuf->page_count = 0; in deactivate_buffer() 845 gobuf->page_count = GO7007_BUF_PAGES + 1; in vidioc_qbuf() 847 gobuf->page_count = GO7007_BUF_PAGES; in vidioc_qbuf() 852 gobuf->user_addr & PAGE_MASK, gobuf->page_count, in vidioc_qbuf() 856 if (ret != gobuf->page_count) { in vidioc_qbuf() 860 gobuf->page_count = 0; in vidioc_qbuf()
|
/drivers/net/mlx4/ |
D | icm.h | 131 int mlx4_UNMAP_ICM(struct mlx4_dev *dev, u64 virt, u32 page_count);
|
D | icm.c | 211 int mlx4_UNMAP_ICM(struct mlx4_dev *dev, u64 virt, u32 page_count) in mlx4_UNMAP_ICM() argument 213 return mlx4_cmd(dev, virt, page_count, 0, MLX4_CMD_UNMAP_ICM, in mlx4_UNMAP_ICM()
|