/drivers/media/platform/soc_camera/ |
D | soc_mediabus.c | 26 .order = SOC_MBUS_ORDER_LE, 36 .order = SOC_MBUS_ORDER_LE, 46 .order = SOC_MBUS_ORDER_LE, 56 .order = SOC_MBUS_ORDER_LE, 66 .order = SOC_MBUS_ORDER_LE, 76 .order = SOC_MBUS_ORDER_BE, 86 .order = SOC_MBUS_ORDER_LE, 96 .order = SOC_MBUS_ORDER_BE, 106 .order = SOC_MBUS_ORDER_LE, 115 .order = SOC_MBUS_ORDER_LE, [all …]
|
/drivers/media/pci/cx18/ |
D | cx18-mailbox.c | 244 static void epu_dma_done(struct cx18 *cx, struct cx18_in_work_order *order) in epu_dma_done() argument 253 mb = &order->mb; in epu_dma_done() 260 (order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) ? in epu_dma_done() 266 mdl_ack = order->mdl_ack; in epu_dma_done() 290 if ((order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) && in epu_dma_done() 338 static void epu_debug(struct cx18 *cx, struct cx18_in_work_order *order) in epu_debug() argument 341 char *str = order->str; in epu_debug() 343 CX18_DEBUG_INFO("%x %s\n", order->mb.args[0], str); in epu_debug() 349 static void epu_cmd(struct cx18 *cx, struct cx18_in_work_order *order) in epu_cmd() argument 351 switch (order->rpu) { in epu_cmd() [all …]
|
/drivers/staging/android/ion/ |
D | ion_system_heap.c | 34 static int order_to_index(unsigned int order) in order_to_index() argument 39 if (order == orders[i]) in order_to_index() 45 static inline unsigned int order_to_size(int order) in order_to_size() argument 47 return PAGE_SIZE << order; in order_to_size() 57 unsigned long order) in alloc_buffer_page() argument 60 struct ion_page_pool *pool = heap->pools[order_to_index(order)]; in alloc_buffer_page() 68 if (order > 4) in alloc_buffer_page() 70 page = alloc_pages(gfp_flags | __GFP_COMP, order); in alloc_buffer_page() 73 ion_pages_sync_for_device(NULL, page, PAGE_SIZE << order, in alloc_buffer_page() 83 unsigned int order = compound_order(page); in free_buffer_page() local [all …]
|
D | ion_page_pool.c | 29 struct page *page = alloc_pages(pool->gfp_mask, pool->order); in ion_page_pool_alloc_pages() 35 ion_pages_sync_for_device(NULL, page, PAGE_SIZE << pool->order, in ion_page_pool_alloc_pages() 44 __free_pages(page, pool->order); in ion_page_pool_free_pages() 102 BUG_ON(pool->order != compound_order(page)); in ion_page_pool_free() 121 return count << pool->order; in ion_page_pool_total() 157 struct ion_page_pool *ion_page_pool_create(gfp_t gfp_mask, unsigned int order) in ion_page_pool_create() argument 168 pool->order = order; in ion_page_pool_create() 170 plist_node_init(&pool->list, order); in ion_page_pool_create()
|
D | ion_priv.h | 378 unsigned int order; member 382 struct ion_page_pool *ion_page_pool_create(gfp_t gfp_mask, unsigned int order); 395 set_memory_wc((unsigned long)va, 1 << pool->order); in ion_page_pool_alloc_set_cache_policy() 404 set_memory_wb((unsigned long)va, 1 << pool->order); in ion_page_pool_free_set_cache_policy()
|
/drivers/infiniband/hw/ehca/ |
D | ipz_pt_fn.c | 129 int order = ilog2(queue->pagesize) - 9; in alloc_small_queue_page() local 135 if (!list_empty(&pd->free[order])) in alloc_small_queue_page() 136 page = list_entry(pd->free[order].next, in alloc_small_queue_page() 149 list_add(&page->list, &pd->free[order]); in alloc_small_queue_page() 152 bit = find_first_zero_bit(page->bitmap, IPZ_SPAGE_PER_KPAGE >> order); in alloc_small_queue_page() 156 if (page->fill == IPZ_SPAGE_PER_KPAGE >> order) in alloc_small_queue_page() 157 list_move(&page->list, &pd->full[order]); in alloc_small_queue_page() 161 queue->queue_pages[0] = (void *)(page->page | (bit << (order + 9))); in alloc_small_queue_page() 163 queue->offset = bit << (order + 9); in alloc_small_queue_page() 174 int order = ilog2(queue->pagesize) - 9; in free_small_queue_page() local [all …]
|
/drivers/net/ethernet/mellanox/mlx4/ |
D | mr.c | 46 static u32 mlx4_buddy_alloc(struct mlx4_buddy *buddy, int order) in mlx4_buddy_alloc() argument 54 for (o = order; o <= buddy->max_order; ++o) in mlx4_buddy_alloc() 69 while (o > order) { in mlx4_buddy_alloc() 78 seg <<= order; in mlx4_buddy_alloc() 83 static void mlx4_buddy_free(struct mlx4_buddy *buddy, u32 seg, int order) in mlx4_buddy_free() argument 85 seg >>= order; in mlx4_buddy_free() 89 while (test_bit(seg ^ 1, buddy->bits[order])) { in mlx4_buddy_free() 90 clear_bit(seg ^ 1, buddy->bits[order]); in mlx4_buddy_free() 91 --buddy->num_free[order]; in mlx4_buddy_free() 93 ++order; in mlx4_buddy_free() [all …]
|
D | alloc.c | 286 struct mlx4_db *db, int order) in mlx4_alloc_db_from_pgdir() argument 291 for (o = order; o <= 1; ++o) { in mlx4_alloc_db_from_pgdir() 304 if (o > order) in mlx4_alloc_db_from_pgdir() 305 set_bit(i ^ 1, pgdir->bits[order]); in mlx4_alloc_db_from_pgdir() 311 db->order = order; in mlx4_alloc_db_from_pgdir() 316 int mlx4_db_alloc(struct mlx4_dev *dev, struct mlx4_db *db, int order, gfp_t gfp) in mlx4_db_alloc() argument 325 if (!mlx4_alloc_db_from_pgdir(pgdir, db, order)) in mlx4_db_alloc() 337 WARN_ON(mlx4_alloc_db_from_pgdir(pgdir, db, order)); in mlx4_db_alloc() 354 o = db->order; in mlx4_db_free() 357 if (db->order == 0 && test_bit(i ^ 1, db->u.pgdir->order0)) { in mlx4_db_free()
|
D | icm.c | 96 static int mlx4_alloc_icm_pages(struct scatterlist *mem, int order, in mlx4_alloc_icm_pages() argument 101 page = alloc_pages_node(node, gfp_mask, order); in mlx4_alloc_icm_pages() 103 page = alloc_pages(gfp_mask, order); in mlx4_alloc_icm_pages() 108 sg_set_page(mem, page, PAGE_SIZE << order, 0); in mlx4_alloc_icm_pages() 113 int order, gfp_t gfp_mask) in mlx4_alloc_icm_coherent() argument 115 void *buf = dma_alloc_coherent(dev, PAGE_SIZE << order, in mlx4_alloc_icm_coherent() 121 dma_free_coherent(dev, PAGE_SIZE << order, in mlx4_alloc_icm_coherent() 126 sg_set_buf(mem, buf, PAGE_SIZE << order); in mlx4_alloc_icm_coherent() 127 sg_dma_len(mem) = PAGE_SIZE << order; in mlx4_alloc_icm_coherent()
|
/drivers/gpu/drm/ |
D | drm_hashtab.c | 41 int drm_ht_create(struct drm_open_hash *ht, unsigned int order) in drm_ht_create() argument 43 unsigned int size = 1 << order; in drm_ht_create() 45 ht->order = order; in drm_ht_create() 66 hashed_key = hash_long(key, ht->order); in drm_ht_verbose_list() 80 hashed_key = hash_long(key, ht->order); in drm_ht_find_key() 98 hashed_key = hash_long(key, ht->order); in drm_ht_find_key_rcu() 117 hashed_key = hash_long(key, ht->order); in drm_ht_insert_item() 201 if ((PAGE_SIZE / sizeof(*ht->table)) >> ht->order) in drm_ht_remove()
|
D | drm_bufs.c | 607 int order; in drm_legacy_addbufs_agp() local 620 order = order_base_2(request->size); in drm_legacy_addbufs_agp() 621 size = 1 << order; in drm_legacy_addbufs_agp() 625 page_order = order - PAGE_SHIFT > 0 ? order - PAGE_SHIFT : 0; in drm_legacy_addbufs_agp() 632 DRM_DEBUG("order: %d\n", order); in drm_legacy_addbufs_agp() 639 if (order < DRM_MIN_ORDER || order > DRM_MAX_ORDER) in drm_legacy_addbufs_agp() 664 entry = &dma->bufs[order]; in drm_legacy_addbufs_agp() 693 buf->order = order; in drm_legacy_addbufs_agp() 766 int order; in drm_legacy_addbufs_pci() local 791 order = order_base_2(request->size); in drm_legacy_addbufs_pci() [all …]
|
/drivers/media/v4l2-core/ |
D | videobuf2-dma-sg.c | 56 int order; in vb2_dma_sg_alloc_compacted() local 59 order = get_order(size); in vb2_dma_sg_alloc_compacted() 61 if ((PAGE_SIZE << order) > size) in vb2_dma_sg_alloc_compacted() 62 order--; in vb2_dma_sg_alloc_compacted() 67 __GFP_NOWARN | gfp_flags, order); in vb2_dma_sg_alloc_compacted() 71 if (order == 0) { in vb2_dma_sg_alloc_compacted() 76 order--; in vb2_dma_sg_alloc_compacted() 79 split_page(pages, order); in vb2_dma_sg_alloc_compacted() 80 for (i = 0; i < (1 << order); i++) in vb2_dma_sg_alloc_compacted() 83 size -= PAGE_SIZE << order; in vb2_dma_sg_alloc_compacted()
|
/drivers/xen/ |
D | swiotlb-xen.c | 220 unsigned long bytes, order; in xen_swiotlb_init() local 228 order = get_order(xen_io_tlb_nslabs << IO_TLB_SHIFT); in xen_swiotlb_init() 237 while ((SLABS_PER_PAGE << order) > IO_TLB_MIN_SLABS) { in xen_swiotlb_init() 238 xen_io_tlb_start = (void *)__get_free_pages(__GFP_NOWARN, order); in xen_swiotlb_init() 241 order--; in xen_swiotlb_init() 243 if (order != get_order(bytes)) { in xen_swiotlb_init() 245 (PAGE_SIZE << order) >> 20); in xen_swiotlb_init() 246 xen_io_tlb_nslabs = SLABS_PER_PAGE << order; in xen_swiotlb_init() 265 free_pages((unsigned long)xen_io_tlb_start, order); in xen_swiotlb_init() 292 free_pages((unsigned long)xen_io_tlb_start, order); in xen_swiotlb_init() [all …]
|
/drivers/infiniband/hw/mlx5/ |
D | mr.c | 57 static int order2idx(struct mlx5_ib_dev *dev, int order) in order2idx() argument 61 if (order < cache->ent[0].order) in order2idx() 64 return order - cache->ent[0].order; in order2idx() 72 int c = order2idx(dev, mr->order); in reg_mr_callback() 127 int npages = 1 << ent->order; in add_keys() 146 mr->order = ent->order; in add_keys() 212 c = order2idx(dev, ent->order); in size_write() 278 c = order2idx(dev, ent->order); in limit_write() 343 int i = order2idx(dev, ent->order); in __cache_work_func() 395 static struct mlx5_ib_mr *alloc_cached_mr(struct mlx5_ib_dev *dev, int order) in alloc_cached_mr() argument [all …]
|
D | mem.c | 45 int *ncont, int *order) in mlx5_ib_cont_pages() argument 95 if (order) in mlx5_ib_cont_pages() 96 *order = ilog2(roundup_pow_of_two(i) >> m); in mlx5_ib_cont_pages() 102 if (order) in mlx5_ib_cont_pages() 103 *order = 0; in mlx5_ib_cont_pages()
|
/drivers/atm/ |
D | eni.c | 173 1 << eni_dev->free_list[i].order); in dump_mem() 205 int len,order; in eni_put_free() local 217 for (order = 0; !(((unsigned long)start | size) & (1 << order)); order++); in eni_put_free() 218 if (MID_MIN_BUF_SIZE > (1 << order)) { in eni_put_free() 220 order); in eni_put_free() 224 list[len].order = order; in eni_put_free() 226 start += 1 << order; in eni_put_free() 227 size -= 1 << order; in eni_put_free() 238 int len,i,order,best_order,index; in eni_alloc_mem() local 244 for (order = 0; (1 << order) < *size; order++); in eni_alloc_mem() [all …]
|
/drivers/infiniband/hw/mthca/ |
D | mthca_mr.c | 43 int order; member 84 static u32 mthca_buddy_alloc(struct mthca_buddy *buddy, int order) in mthca_buddy_alloc() argument 92 for (o = order; o <= buddy->max_order; ++o) in mthca_buddy_alloc() 107 while (o > order) { in mthca_buddy_alloc() 116 seg <<= order; in mthca_buddy_alloc() 121 static void mthca_buddy_free(struct mthca_buddy *buddy, u32 seg, int order) in mthca_buddy_free() argument 123 seg >>= order; in mthca_buddy_free() 127 while (test_bit(seg ^ 1, buddy->bits[order])) { in mthca_buddy_free() 128 clear_bit(seg ^ 1, buddy->bits[order]); in mthca_buddy_free() 129 --buddy->num_free[order]; in mthca_buddy_free() [all …]
|
/drivers/char/agp/ |
D | amd64-agp.c | 273 int order = 0; in fix_northbridge() local 295 order = 7 - hweight16(apsize); in fix_northbridge() 303 if (order >=0 && aper + (32ULL<<(20 + order)) > 0x100000000ULL) { in fix_northbridge() 305 32 << order); in fix_northbridge() 306 order = nb_order; in fix_northbridge() 309 if (nb_order >= order) { in fix_northbridge() 315 aper, 32 << order); in fix_northbridge() 316 if (order < 0 || !agp_aperture_valid(aper, (32*1024*1024)<<order)) in fix_northbridge() 319 gart_set_size_and_enable(nb, order); in fix_northbridge()
|
/drivers/staging/comedi/drivers/ |
D | addi_apci_3120.c | 60 int ret, order, i; in apci3120_auto_attach() local 93 for (order = 2; order >= 0; order--) { in apci3120_auto_attach() 95 dma_alloc_coherent(dev->hw_dev, PAGE_SIZE << order, in apci3120_auto_attach() 104 devpriv->ui_DmaBufferSize[i] = PAGE_SIZE << order; in apci3120_auto_attach()
|
/drivers/s390/char/ |
D | tape_3590.c | 204 struct tape3592_kekl_query_order *order; in tape_3592_kekl_query() local 212 request = tape_alloc_request(2, sizeof(*order)); in tape_3592_kekl_query() 217 order = request->cpdata; in tape_3592_kekl_query() 218 memset(order,0,sizeof(*order)); in tape_3592_kekl_query() 219 order->code = 0xe2; in tape_3592_kekl_query() 220 order->max_count = 2; in tape_3592_kekl_query() 222 tape_ccw_cc(request->cpaddr, PERF_SUBSYS_FUNC, sizeof(*order), order); in tape_3592_kekl_query() 277 struct tape3592_kekl_set_order *order; in tape_3592_kekl_set() local 286 request = tape_alloc_request(1, sizeof(*order)); in tape_3592_kekl_set() 289 order = request->cpdata; in tape_3592_kekl_set() [all …]
|
D | hmcdrv_ftp.c | 239 int order; in hmcdrv_ftp_cmd() local 247 order = get_order(ftp.len); in hmcdrv_ftp_cmd() 248 ftp.buf = (void *) __get_free_pages(GFP_KERNEL | GFP_DMA, order); in hmcdrv_ftp_cmd() 281 free_pages((unsigned long) ftp.buf, order); in hmcdrv_ftp_cmd()
|
/drivers/gpu/drm/nouveau/core/subdev/fb/ |
D | ramgk20a.c | 59 u32 npages, order; in gk20a_ram_get() local 74 order = fls(align); in gk20a_ram_get() 76 order--; in gk20a_ram_get() 77 align = BIT(order); in gk20a_ram_get()
|
/drivers/media/pci/solo6x10/ |
D | solo6x10-p2m.c | 222 int order = get_order(size); in solo_p2m_test() local 224 wr_buf = (u32 *)__get_free_pages(GFP_KERNEL, order); in solo_p2m_test() 228 rd_buf = (u32 *)__get_free_pages(GFP_KERNEL, order); in solo_p2m_test() 230 free_pages((unsigned long)wr_buf, order); in solo_p2m_test() 256 free_pages((unsigned long)wr_buf, order); in solo_p2m_test() 257 free_pages((unsigned long)rd_buf, order); in solo_p2m_test()
|
/drivers/media/platform/omap/ |
D | omap_voutlib.c | 303 u32 order, size; in omap_vout_alloc_buffer() local 307 order = get_order(size); in omap_vout_alloc_buffer() 308 virt_addr = __get_free_pages(GFP_KERNEL, order); in omap_vout_alloc_buffer() 327 u32 order, size; in omap_vout_free_buffer() local 331 order = get_order(size); in omap_vout_free_buffer() 338 free_pages((unsigned long) virtaddr, order); in omap_vout_free_buffer()
|
/drivers/usb/atm/ |
D | Kconfig | 25 modem. In order to use your modem you will need to install the 37 AccessRunner chipset. In order to use your modem you will need to 49 or eagle chipset. In order to use your modem you will need to 60 another USB DSL drivers. In order to use your modem you will need to
|