Home
last modified time | relevance | path

Searched refs:order (Results 1 – 25 of 198) sorted by relevance

12345678

/drivers/media/platform/soc_camera/
Dsoc_mediabus.c26 .order = SOC_MBUS_ORDER_LE,
36 .order = SOC_MBUS_ORDER_LE,
46 .order = SOC_MBUS_ORDER_LE,
56 .order = SOC_MBUS_ORDER_LE,
66 .order = SOC_MBUS_ORDER_LE,
76 .order = SOC_MBUS_ORDER_BE,
86 .order = SOC_MBUS_ORDER_LE,
96 .order = SOC_MBUS_ORDER_BE,
106 .order = SOC_MBUS_ORDER_LE,
115 .order = SOC_MBUS_ORDER_LE,
[all …]
/drivers/media/pci/cx18/
Dcx18-mailbox.c244 static void epu_dma_done(struct cx18 *cx, struct cx18_in_work_order *order) in epu_dma_done() argument
253 mb = &order->mb; in epu_dma_done()
260 (order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) ? in epu_dma_done()
266 mdl_ack = order->mdl_ack; in epu_dma_done()
290 if ((order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) && in epu_dma_done()
338 static void epu_debug(struct cx18 *cx, struct cx18_in_work_order *order) in epu_debug() argument
341 char *str = order->str; in epu_debug()
343 CX18_DEBUG_INFO("%x %s\n", order->mb.args[0], str); in epu_debug()
349 static void epu_cmd(struct cx18 *cx, struct cx18_in_work_order *order) in epu_cmd() argument
351 switch (order->rpu) { in epu_cmd()
[all …]
/drivers/staging/android/ion/
Dion_system_heap.c34 static int order_to_index(unsigned int order) in order_to_index() argument
39 if (order == orders[i]) in order_to_index()
45 static inline unsigned int order_to_size(int order) in order_to_size() argument
47 return PAGE_SIZE << order; in order_to_size()
57 unsigned long order) in alloc_buffer_page() argument
60 struct ion_page_pool *pool = heap->pools[order_to_index(order)]; in alloc_buffer_page()
68 if (order > 4) in alloc_buffer_page()
70 page = alloc_pages(gfp_flags | __GFP_COMP, order); in alloc_buffer_page()
73 ion_pages_sync_for_device(NULL, page, PAGE_SIZE << order, in alloc_buffer_page()
83 unsigned int order = compound_order(page); in free_buffer_page() local
[all …]
Dion_page_pool.c29 struct page *page = alloc_pages(pool->gfp_mask, pool->order); in ion_page_pool_alloc_pages()
35 ion_pages_sync_for_device(NULL, page, PAGE_SIZE << pool->order, in ion_page_pool_alloc_pages()
44 __free_pages(page, pool->order); in ion_page_pool_free_pages()
102 BUG_ON(pool->order != compound_order(page)); in ion_page_pool_free()
121 return count << pool->order; in ion_page_pool_total()
157 struct ion_page_pool *ion_page_pool_create(gfp_t gfp_mask, unsigned int order) in ion_page_pool_create() argument
168 pool->order = order; in ion_page_pool_create()
170 plist_node_init(&pool->list, order); in ion_page_pool_create()
Dion_priv.h378 unsigned int order; member
382 struct ion_page_pool *ion_page_pool_create(gfp_t gfp_mask, unsigned int order);
395 set_memory_wc((unsigned long)va, 1 << pool->order); in ion_page_pool_alloc_set_cache_policy()
404 set_memory_wb((unsigned long)va, 1 << pool->order); in ion_page_pool_free_set_cache_policy()
/drivers/infiniband/hw/ehca/
Dipz_pt_fn.c129 int order = ilog2(queue->pagesize) - 9; in alloc_small_queue_page() local
135 if (!list_empty(&pd->free[order])) in alloc_small_queue_page()
136 page = list_entry(pd->free[order].next, in alloc_small_queue_page()
149 list_add(&page->list, &pd->free[order]); in alloc_small_queue_page()
152 bit = find_first_zero_bit(page->bitmap, IPZ_SPAGE_PER_KPAGE >> order); in alloc_small_queue_page()
156 if (page->fill == IPZ_SPAGE_PER_KPAGE >> order) in alloc_small_queue_page()
157 list_move(&page->list, &pd->full[order]); in alloc_small_queue_page()
161 queue->queue_pages[0] = (void *)(page->page | (bit << (order + 9))); in alloc_small_queue_page()
163 queue->offset = bit << (order + 9); in alloc_small_queue_page()
174 int order = ilog2(queue->pagesize) - 9; in free_small_queue_page() local
[all …]
/drivers/net/ethernet/mellanox/mlx4/
Dmr.c46 static u32 mlx4_buddy_alloc(struct mlx4_buddy *buddy, int order) in mlx4_buddy_alloc() argument
54 for (o = order; o <= buddy->max_order; ++o) in mlx4_buddy_alloc()
69 while (o > order) { in mlx4_buddy_alloc()
78 seg <<= order; in mlx4_buddy_alloc()
83 static void mlx4_buddy_free(struct mlx4_buddy *buddy, u32 seg, int order) in mlx4_buddy_free() argument
85 seg >>= order; in mlx4_buddy_free()
89 while (test_bit(seg ^ 1, buddy->bits[order])) { in mlx4_buddy_free()
90 clear_bit(seg ^ 1, buddy->bits[order]); in mlx4_buddy_free()
91 --buddy->num_free[order]; in mlx4_buddy_free()
93 ++order; in mlx4_buddy_free()
[all …]
Dalloc.c286 struct mlx4_db *db, int order) in mlx4_alloc_db_from_pgdir() argument
291 for (o = order; o <= 1; ++o) { in mlx4_alloc_db_from_pgdir()
304 if (o > order) in mlx4_alloc_db_from_pgdir()
305 set_bit(i ^ 1, pgdir->bits[order]); in mlx4_alloc_db_from_pgdir()
311 db->order = order; in mlx4_alloc_db_from_pgdir()
316 int mlx4_db_alloc(struct mlx4_dev *dev, struct mlx4_db *db, int order, gfp_t gfp) in mlx4_db_alloc() argument
325 if (!mlx4_alloc_db_from_pgdir(pgdir, db, order)) in mlx4_db_alloc()
337 WARN_ON(mlx4_alloc_db_from_pgdir(pgdir, db, order)); in mlx4_db_alloc()
354 o = db->order; in mlx4_db_free()
357 if (db->order == 0 && test_bit(i ^ 1, db->u.pgdir->order0)) { in mlx4_db_free()
Dicm.c96 static int mlx4_alloc_icm_pages(struct scatterlist *mem, int order, in mlx4_alloc_icm_pages() argument
101 page = alloc_pages_node(node, gfp_mask, order); in mlx4_alloc_icm_pages()
103 page = alloc_pages(gfp_mask, order); in mlx4_alloc_icm_pages()
108 sg_set_page(mem, page, PAGE_SIZE << order, 0); in mlx4_alloc_icm_pages()
113 int order, gfp_t gfp_mask) in mlx4_alloc_icm_coherent() argument
115 void *buf = dma_alloc_coherent(dev, PAGE_SIZE << order, in mlx4_alloc_icm_coherent()
121 dma_free_coherent(dev, PAGE_SIZE << order, in mlx4_alloc_icm_coherent()
126 sg_set_buf(mem, buf, PAGE_SIZE << order); in mlx4_alloc_icm_coherent()
127 sg_dma_len(mem) = PAGE_SIZE << order; in mlx4_alloc_icm_coherent()
/drivers/gpu/drm/
Ddrm_hashtab.c41 int drm_ht_create(struct drm_open_hash *ht, unsigned int order) in drm_ht_create() argument
43 unsigned int size = 1 << order; in drm_ht_create()
45 ht->order = order; in drm_ht_create()
66 hashed_key = hash_long(key, ht->order); in drm_ht_verbose_list()
80 hashed_key = hash_long(key, ht->order); in drm_ht_find_key()
98 hashed_key = hash_long(key, ht->order); in drm_ht_find_key_rcu()
117 hashed_key = hash_long(key, ht->order); in drm_ht_insert_item()
201 if ((PAGE_SIZE / sizeof(*ht->table)) >> ht->order) in drm_ht_remove()
Ddrm_bufs.c607 int order; in drm_legacy_addbufs_agp() local
620 order = order_base_2(request->size); in drm_legacy_addbufs_agp()
621 size = 1 << order; in drm_legacy_addbufs_agp()
625 page_order = order - PAGE_SHIFT > 0 ? order - PAGE_SHIFT : 0; in drm_legacy_addbufs_agp()
632 DRM_DEBUG("order: %d\n", order); in drm_legacy_addbufs_agp()
639 if (order < DRM_MIN_ORDER || order > DRM_MAX_ORDER) in drm_legacy_addbufs_agp()
664 entry = &dma->bufs[order]; in drm_legacy_addbufs_agp()
693 buf->order = order; in drm_legacy_addbufs_agp()
766 int order; in drm_legacy_addbufs_pci() local
791 order = order_base_2(request->size); in drm_legacy_addbufs_pci()
[all …]
/drivers/media/v4l2-core/
Dvideobuf2-dma-sg.c56 int order; in vb2_dma_sg_alloc_compacted() local
59 order = get_order(size); in vb2_dma_sg_alloc_compacted()
61 if ((PAGE_SIZE << order) > size) in vb2_dma_sg_alloc_compacted()
62 order--; in vb2_dma_sg_alloc_compacted()
67 __GFP_NOWARN | gfp_flags, order); in vb2_dma_sg_alloc_compacted()
71 if (order == 0) { in vb2_dma_sg_alloc_compacted()
76 order--; in vb2_dma_sg_alloc_compacted()
79 split_page(pages, order); in vb2_dma_sg_alloc_compacted()
80 for (i = 0; i < (1 << order); i++) in vb2_dma_sg_alloc_compacted()
83 size -= PAGE_SIZE << order; in vb2_dma_sg_alloc_compacted()
/drivers/xen/
Dswiotlb-xen.c220 unsigned long bytes, order; in xen_swiotlb_init() local
228 order = get_order(xen_io_tlb_nslabs << IO_TLB_SHIFT); in xen_swiotlb_init()
237 while ((SLABS_PER_PAGE << order) > IO_TLB_MIN_SLABS) { in xen_swiotlb_init()
238 xen_io_tlb_start = (void *)__get_free_pages(__GFP_NOWARN, order); in xen_swiotlb_init()
241 order--; in xen_swiotlb_init()
243 if (order != get_order(bytes)) { in xen_swiotlb_init()
245 (PAGE_SIZE << order) >> 20); in xen_swiotlb_init()
246 xen_io_tlb_nslabs = SLABS_PER_PAGE << order; in xen_swiotlb_init()
265 free_pages((unsigned long)xen_io_tlb_start, order); in xen_swiotlb_init()
292 free_pages((unsigned long)xen_io_tlb_start, order); in xen_swiotlb_init()
[all …]
/drivers/infiniband/hw/mlx5/
Dmr.c57 static int order2idx(struct mlx5_ib_dev *dev, int order) in order2idx() argument
61 if (order < cache->ent[0].order) in order2idx()
64 return order - cache->ent[0].order; in order2idx()
72 int c = order2idx(dev, mr->order); in reg_mr_callback()
127 int npages = 1 << ent->order; in add_keys()
146 mr->order = ent->order; in add_keys()
212 c = order2idx(dev, ent->order); in size_write()
278 c = order2idx(dev, ent->order); in limit_write()
343 int i = order2idx(dev, ent->order); in __cache_work_func()
395 static struct mlx5_ib_mr *alloc_cached_mr(struct mlx5_ib_dev *dev, int order) in alloc_cached_mr() argument
[all …]
Dmem.c45 int *ncont, int *order) in mlx5_ib_cont_pages() argument
95 if (order) in mlx5_ib_cont_pages()
96 *order = ilog2(roundup_pow_of_two(i) >> m); in mlx5_ib_cont_pages()
102 if (order) in mlx5_ib_cont_pages()
103 *order = 0; in mlx5_ib_cont_pages()
/drivers/atm/
Deni.c173 1 << eni_dev->free_list[i].order); in dump_mem()
205 int len,order; in eni_put_free() local
217 for (order = 0; !(((unsigned long)start | size) & (1 << order)); order++); in eni_put_free()
218 if (MID_MIN_BUF_SIZE > (1 << order)) { in eni_put_free()
220 order); in eni_put_free()
224 list[len].order = order; in eni_put_free()
226 start += 1 << order; in eni_put_free()
227 size -= 1 << order; in eni_put_free()
238 int len,i,order,best_order,index; in eni_alloc_mem() local
244 for (order = 0; (1 << order) < *size; order++); in eni_alloc_mem()
[all …]
/drivers/infiniband/hw/mthca/
Dmthca_mr.c43 int order; member
84 static u32 mthca_buddy_alloc(struct mthca_buddy *buddy, int order) in mthca_buddy_alloc() argument
92 for (o = order; o <= buddy->max_order; ++o) in mthca_buddy_alloc()
107 while (o > order) { in mthca_buddy_alloc()
116 seg <<= order; in mthca_buddy_alloc()
121 static void mthca_buddy_free(struct mthca_buddy *buddy, u32 seg, int order) in mthca_buddy_free() argument
123 seg >>= order; in mthca_buddy_free()
127 while (test_bit(seg ^ 1, buddy->bits[order])) { in mthca_buddy_free()
128 clear_bit(seg ^ 1, buddy->bits[order]); in mthca_buddy_free()
129 --buddy->num_free[order]; in mthca_buddy_free()
[all …]
/drivers/char/agp/
Damd64-agp.c273 int order = 0; in fix_northbridge() local
295 order = 7 - hweight16(apsize); in fix_northbridge()
303 if (order >=0 && aper + (32ULL<<(20 + order)) > 0x100000000ULL) { in fix_northbridge()
305 32 << order); in fix_northbridge()
306 order = nb_order; in fix_northbridge()
309 if (nb_order >= order) { in fix_northbridge()
315 aper, 32 << order); in fix_northbridge()
316 if (order < 0 || !agp_aperture_valid(aper, (32*1024*1024)<<order)) in fix_northbridge()
319 gart_set_size_and_enable(nb, order); in fix_northbridge()
/drivers/staging/comedi/drivers/
Daddi_apci_3120.c60 int ret, order, i; in apci3120_auto_attach() local
93 for (order = 2; order >= 0; order--) { in apci3120_auto_attach()
95 dma_alloc_coherent(dev->hw_dev, PAGE_SIZE << order, in apci3120_auto_attach()
104 devpriv->ui_DmaBufferSize[i] = PAGE_SIZE << order; in apci3120_auto_attach()
/drivers/s390/char/
Dtape_3590.c204 struct tape3592_kekl_query_order *order; in tape_3592_kekl_query() local
212 request = tape_alloc_request(2, sizeof(*order)); in tape_3592_kekl_query()
217 order = request->cpdata; in tape_3592_kekl_query()
218 memset(order,0,sizeof(*order)); in tape_3592_kekl_query()
219 order->code = 0xe2; in tape_3592_kekl_query()
220 order->max_count = 2; in tape_3592_kekl_query()
222 tape_ccw_cc(request->cpaddr, PERF_SUBSYS_FUNC, sizeof(*order), order); in tape_3592_kekl_query()
277 struct tape3592_kekl_set_order *order; in tape_3592_kekl_set() local
286 request = tape_alloc_request(1, sizeof(*order)); in tape_3592_kekl_set()
289 order = request->cpdata; in tape_3592_kekl_set()
[all …]
Dhmcdrv_ftp.c239 int order; in hmcdrv_ftp_cmd() local
247 order = get_order(ftp.len); in hmcdrv_ftp_cmd()
248 ftp.buf = (void *) __get_free_pages(GFP_KERNEL | GFP_DMA, order); in hmcdrv_ftp_cmd()
281 free_pages((unsigned long) ftp.buf, order); in hmcdrv_ftp_cmd()
/drivers/gpu/drm/nouveau/core/subdev/fb/
Dramgk20a.c59 u32 npages, order; in gk20a_ram_get() local
74 order = fls(align); in gk20a_ram_get()
76 order--; in gk20a_ram_get()
77 align = BIT(order); in gk20a_ram_get()
/drivers/media/pci/solo6x10/
Dsolo6x10-p2m.c222 int order = get_order(size); in solo_p2m_test() local
224 wr_buf = (u32 *)__get_free_pages(GFP_KERNEL, order); in solo_p2m_test()
228 rd_buf = (u32 *)__get_free_pages(GFP_KERNEL, order); in solo_p2m_test()
230 free_pages((unsigned long)wr_buf, order); in solo_p2m_test()
256 free_pages((unsigned long)wr_buf, order); in solo_p2m_test()
257 free_pages((unsigned long)rd_buf, order); in solo_p2m_test()
/drivers/media/platform/omap/
Domap_voutlib.c303 u32 order, size; in omap_vout_alloc_buffer() local
307 order = get_order(size); in omap_vout_alloc_buffer()
308 virt_addr = __get_free_pages(GFP_KERNEL, order); in omap_vout_alloc_buffer()
327 u32 order, size; in omap_vout_free_buffer() local
331 order = get_order(size); in omap_vout_free_buffer()
338 free_pages((unsigned long) virtaddr, order); in omap_vout_free_buffer()
/drivers/usb/atm/
DKconfig25 modem. In order to use your modem you will need to install the
37 AccessRunner chipset. In order to use your modem you will need to
49 or eagle chipset. In order to use your modem you will need to
60 another USB DSL drivers. In order to use your modem you will need to

12345678