Home
last modified time | relevance | path

Searched refs:chunk_size (Results 1 – 25 of 50) sorted by relevance

12

/drivers/md/
Ddm-exception-store.c145 unsigned chunk_size; in set_chunk_size() local
147 if (kstrtouint(chunk_size_arg, 10, &chunk_size)) { in set_chunk_size()
152 if (!chunk_size) { in set_chunk_size()
153 store->chunk_size = store->chunk_mask = store->chunk_shift = 0; in set_chunk_size()
157 return dm_exception_store_set_chunk_size(store, chunk_size, error); in set_chunk_size()
161 unsigned chunk_size, in dm_exception_store_set_chunk_size() argument
165 if (!is_power_of_2(chunk_size)) { in dm_exception_store_set_chunk_size()
171 if (chunk_size % in dm_exception_store_set_chunk_size()
173 chunk_size % in dm_exception_store_set_chunk_size()
179 if (chunk_size > INT_MAX >> SECTOR_SHIFT) { in dm_exception_store_set_chunk_size()
[all …]
Ddm-stripe.c35 uint32_t chunk_size; member
103 uint32_t chunk_size; in stripe_ctr() local
117 if (kstrtouint(argv[1], 10, &chunk_size) || !chunk_size) { in stripe_ctr()
130 if (sector_div(tmp_len, chunk_size)) { in stripe_ctr()
164 r = dm_set_target_max_io_len(ti, chunk_size); in stripe_ctr()
175 sc->chunk_size = chunk_size; in stripe_ctr()
176 if (chunk_size & (chunk_size - 1)) in stripe_ctr()
179 sc->chunk_size_shift = __ffs(chunk_size); in stripe_ctr()
222 chunk_offset = sector_div(chunk, sc->chunk_size); in stripe_map_sector()
224 chunk_offset = chunk & (sc->chunk_size - 1); in stripe_map_sector()
[all …]
Ddm-snap-persistent.c81 __le32 chunk_size; member
172 len = ps->store->chunk_size << SECTOR_SHIFT; in alloc_area()
234 .sector = ps->store->chunk_size * chunk, in chunk_io()
235 .count = ps->store->chunk_size, in chunk_io()
301 memset(ps->area, 0, ps->store->chunk_size << SECTOR_SHIFT); in zero_memory_area()
314 unsigned chunk_size; in read_header() local
322 if (!ps->store->chunk_size) { in read_header()
323 ps->store->chunk_size = max(DM_CHUNK_SIZE_DEFAULT_SECTORS, in read_header()
326 ps->store->chunk_mask = ps->store->chunk_size - 1; in read_header()
327 ps->store->chunk_shift = __ffs(ps->store->chunk_size); in read_header()
[all …]
Ddm-snap-transient.c45 if (size < (tc->next_free + store->chunk_size)) in transient_prepare_exception()
49 tc->next_free += store->chunk_size; in transient_prepare_exception()
97 DMEMIT(" N %llu", (unsigned long long)store->chunk_size); in transient_status()
Ddm-exception-store.h120 unsigned chunk_size; member
207 unsigned chunk_size,
Ddm-snap.c530 if (l->store->chunk_size < s->store->chunk_size) in __insert_snapshot()
791 unsigned chunk_size = 0; in __minimum_chunk_size() local
795 chunk_size = min_not_zero(chunk_size, in __minimum_chunk_size()
796 snap->store->chunk_size); in __minimum_chunk_size()
798 return (uint32_t) chunk_size; in __minimum_chunk_size()
951 sector_t sector, unsigned chunk_size);
1017 io_size = linear_chunks * s->store->chunk_size; in snapshot_merge_next_chunks()
1251 s->store->chunk_size = 0; in snapshot_ctr()
1265 if (!s->store->chunk_size) { in snapshot_ctr()
1270 r = dm_set_target_max_io_len(ti, s->store->chunk_size); in snapshot_ctr()
[all …]
/drivers/staging/lustre/lustre/obdclass/
Dllog.c149 int chunk_size = handle->lgh_ctxt->loc_chunk_size; in llog_init_handle() local
156 LASSERT(chunk_size >= LLOG_MIN_CHUNK_SIZE); in llog_init_handle()
161 handle->lgh_hdr_size = chunk_size; in llog_init_handle()
229 int chunk_size; in llog_process_thread() local
238 chunk_size = llh->llh_hdr.lrh_len; in llog_process_thread()
240 LASSERT(is_power_of_2(chunk_size)); in llog_process_thread()
242 buf = libcfs_kvzalloc(chunk_size, GFP_NOFS); in llog_process_thread()
275 memset(buf, 0, chunk_size); in llog_process_thread()
277 index, &cur_offset, buf, chunk_size); in llog_process_thread()
288 if (do_div(tmp_offset, chunk_size)) { in llog_process_thread()
[all …]
/drivers/staging/android/ion/
Dion_chunk_heap.c30 unsigned long chunk_size; member
48 allocated_size = ALIGN(size, chunk_heap->chunk_size); in ion_chunk_heap_allocate()
49 num_chunks = allocated_size / chunk_heap->chunk_size; in ion_chunk_heap_allocate()
66 chunk_heap->chunk_size); in ion_chunk_heap_allocate()
70 chunk_heap->chunk_size, 0); in ion_chunk_heap_allocate()
99 allocated_size = ALIGN(buffer->size, chunk_heap->chunk_size); in ion_chunk_heap_free()
138 chunk_heap->chunk_size = (unsigned long)heap_data->priv; in ion_chunk_heap_create()
139 chunk_heap->pool = gen_pool_create(get_order(chunk_heap->chunk_size) + in ion_chunk_heap_create()
/drivers/net/ethernet/mellanox/mlxsw/
Di2c.c285 int off = mlxsw_i2c->cmd.mb_off_in, chunk_size, i, j; in mlxsw_i2c_write() local
292 chunk_size = (in_mbox_size > MLXSW_I2C_BLK_MAX) ? in mlxsw_i2c_write()
294 write_tran.len = MLXSW_I2C_ADDR_WIDTH + chunk_size; in mlxsw_i2c_write()
297 MLXSW_I2C_BLK_MAX * i, chunk_size); in mlxsw_i2c_write()
316 off += chunk_size; in mlxsw_i2c_write()
317 in_mbox_size -= chunk_size; in mlxsw_i2c_write()
353 int num, chunk_size, reg_size, i, j; in mlxsw_i2c_cmd() local
385 chunk_size = (reg_size > MLXSW_I2C_BLK_MAX) ? in mlxsw_i2c_cmd()
387 read_tran[1].len = chunk_size; in mlxsw_i2c_cmd()
409 off += chunk_size; in mlxsw_i2c_cmd()
[all …]
/drivers/media/rc/
Dlirc_dev.c45 unsigned int chunk_size; member
79 unsigned int chunk_size; in lirc_allocate_buffer() local
86 chunk_size = d->chunk_size ? d->chunk_size : bytes_in_key; in lirc_allocate_buffer()
98 err = lirc_buffer_init(ir->buf, chunk_size, buffer_size); in lirc_allocate_buffer()
108 ir->chunk_size = ir->buf->chunk_size; in lirc_allocate_buffer()
491 buf = kzalloc(ir->chunk_size, GFP_KERNEL); in lirc_dev_fop_read()
504 if (length % ir->chunk_size) { in lirc_dev_fop_read()
556 ir->buf->chunk_size); in lirc_dev_fop_read()
558 written += ir->buf->chunk_size; in lirc_dev_fop_read()
/drivers/gpu/drm/amd/amdkfd/
Dkfd_device.c92 unsigned int chunk_size);
403 unsigned int chunk_size) in kfd_gtt_sa_init() argument
407 if (WARN_ON(buf_size < chunk_size)) in kfd_gtt_sa_init()
411 if (WARN_ON(chunk_size == 0)) in kfd_gtt_sa_init()
414 kfd->gtt_sa_chunk_size = chunk_size; in kfd_gtt_sa_init()
415 kfd->gtt_sa_num_of_chunks = buf_size / chunk_size; in kfd_gtt_sa_init()
442 unsigned int chunk_size) in kfd_gtt_sa_calc_gpu_addr() argument
444 return start_addr + bit_num * chunk_size; in kfd_gtt_sa_calc_gpu_addr()
449 unsigned int chunk_size) in kfd_gtt_sa_calc_cpu_addr() argument
451 return (uint32_t *) ((uint64_t) start_addr + bit_num * chunk_size); in kfd_gtt_sa_calc_cpu_addr()
/drivers/net/wireless/marvell/libertas/
Dif_sdio.c448 u32 chunk_size; in if_sdio_prog_helper() local
477 chunk_size = min_t(size_t, size, 60); in if_sdio_prog_helper()
479 *((__le32*)chunk_buffer) = cpu_to_le32(chunk_size); in if_sdio_prog_helper()
480 memcpy(chunk_buffer + 4, firmware, chunk_size); in if_sdio_prog_helper()
489 firmware += chunk_size; in if_sdio_prog_helper()
490 size -= chunk_size; in if_sdio_prog_helper()
544 u32 chunk_size; in if_sdio_prog_real() local
615 chunk_size = min_t(size_t, req_size, 512); in if_sdio_prog_real()
617 memcpy(chunk_buffer, firmware, chunk_size); in if_sdio_prog_real()
623 chunk_buffer, roundup(chunk_size, 32)); in if_sdio_prog_real()
[all …]
/drivers/net/wireless/ath/wcn36xx/
Ddxe.c242 int i, chunk_size = pool->chunk_size; in wcn36xx_dxe_init_tx_bd() local
253 bd_phy_addr += chunk_size; in wcn36xx_dxe_init_tx_bd()
254 bd_cpu_addr += chunk_size; in wcn36xx_dxe_init_tx_bd()
554 wcn->mgmt_mem_pool.chunk_size = WCN36XX_BD_CHUNK_SIZE + in wcn36xx_dxe_allocate_mem_pools()
557 s = wcn->mgmt_mem_pool.chunk_size * WCN36XX_DXE_CH_DESC_NUMB_TX_H; in wcn36xx_dxe_allocate_mem_pools()
569 wcn->data_mem_pool.chunk_size = WCN36XX_BD_CHUNK_SIZE + in wcn36xx_dxe_allocate_mem_pools()
572 s = wcn->data_mem_pool.chunk_size * WCN36XX_DXE_CH_DESC_NUMB_TX_L; in wcn36xx_dxe_allocate_mem_pools()
592 dma_free_coherent(wcn->dev, wcn->mgmt_mem_pool.chunk_size * in wcn36xx_dxe_free_mem_pools()
598 dma_free_coherent(wcn->dev, wcn->data_mem_pool.chunk_size * in wcn36xx_dxe_free_mem_pools()
/drivers/base/regmap/
Dregmap.c1914 size_t chunk_size = val_bytes; in regmap_bulk_write() local
1918 chunk_size = map->max_raw_write; in regmap_bulk_write()
1919 if (chunk_size % val_bytes) in regmap_bulk_write()
1920 chunk_size -= chunk_size % val_bytes; in regmap_bulk_write()
1921 chunk_count = total_size / chunk_size; in regmap_bulk_write()
1922 chunk_stride *= chunk_size / val_bytes; in regmap_bulk_write()
1930 val + (i * chunk_size), in regmap_bulk_write()
1931 chunk_size); in regmap_bulk_write()
1937 if (!ret && chunk_size * i < total_size) { in regmap_bulk_write()
1939 val + (i * chunk_size), in regmap_bulk_write()
[all …]
/drivers/rpmsg/
Dqcom_glink_native.c790 __le32 chunk_size; in qcom_glink_rx_data() member
793 unsigned int chunk_size; in qcom_glink_rx_data() local
806 chunk_size = le32_to_cpu(hdr.chunk_size); in qcom_glink_rx_data()
809 if (avail < sizeof(hdr) + chunk_size) { in qcom_glink_rx_data()
814 if (WARN(chunk_size % 4, "Incoming data must be word aligned\n")) in qcom_glink_rx_data()
835 intent->data = kmalloc(chunk_size + left_size, in qcom_glink_rx_data()
843 intent->size = chunk_size + left_size; in qcom_glink_rx_data()
865 if (intent->size - intent->offset < chunk_size) { in qcom_glink_rx_data()
873 sizeof(hdr), chunk_size); in qcom_glink_rx_data()
874 intent->offset += chunk_size; in qcom_glink_rx_data()
[all …]
/drivers/mtd/nand/
Dpxa3xx_nand.c236 unsigned int chunk_size; member
332 int nchunks = mtd->writesize / info->chunk_size; in pxa3xx_ooblayout_ecc()
350 int nchunks = mtd->writesize / info->chunk_size; in pxa3xx_ooblayout_free()
365 if (mtd->writesize == 4096 && info->chunk_size == 2048) { in pxa3xx_ooblayout_free()
776 info->sg.length = info->chunk_size; in start_data_dma()
1015 info->step_chunk_size = info->chunk_size; in prepare_set_command()
1066 info->step_chunk_size = info->chunk_size; in prepare_set_command()
1465 info->chunk_size = PAGE_CHUNK_SIZE; in pxa3xx_nand_config_ident()
1498 info->chunk_size = ndcr & NDCR_PAGE_SZ ? 2048 : 512; in pxa3xx_nand_detect_config()
1579 info->chunk_size = 2048; in pxa_ecc_init()
[all …]
/drivers/gpu/drm/vmwgfx/
Dvmwgfx_fifo.c413 uint32_t chunk_size = max - next_cmd; in vmw_fifo_res_copy() local
418 if (bytes < chunk_size) in vmw_fifo_res_copy()
419 chunk_size = bytes; in vmw_fifo_res_copy()
423 memcpy(fifo_mem + (next_cmd >> 2), buffer, chunk_size); in vmw_fifo_res_copy()
424 rest = bytes - chunk_size; in vmw_fifo_res_copy()
426 memcpy(fifo_mem + (min >> 2), buffer + (chunk_size >> 2), rest); in vmw_fifo_res_copy()
/drivers/net/ethernet/mellanox/mlx4/
Dicm.c396 unsigned chunk_size; in mlx4_init_icm_table() local
416 chunk_size = MLX4_TABLE_CHUNK_SIZE; in mlx4_init_icm_table()
418 chunk_size = PAGE_ALIGN(size - in mlx4_init_icm_table()
421 table->icm[i] = mlx4_alloc_icm(dev, chunk_size >> PAGE_SHIFT, in mlx4_init_icm_table()
/drivers/staging/media/lirc/
Dlirc_zilog.c893 if (n % rbuf->chunk_size) { in read()
937 if (rbuf->chunk_size > sizeof(buf)) { in read()
940 rbuf->chunk_size); in read()
945 if (m == rbuf->chunk_size) { in read()
947 rbuf->chunk_size); in read()
948 written += rbuf->chunk_size; in read()
1392 .chunk_size = 2,
1498 ir->l.chunk_size, ir->l.buffer_size); in ir_probe()
/drivers/uwb/i1480/dfu/
Dmac.c172 size_t chunk_size; in i1480_fw_cmp() local
176 chunk_size = size < i1480->buf_size ? size : i1480->buf_size; in i1480_fw_cmp()
177 result = i1480->read(i1480, hdr->address + src_itr, chunk_size); in i1480_fw_cmp()
/drivers/media/dvb-frontends/
Ddrxk.h48 int chunk_size; member
/drivers/gpu/drm/qxl/
Dqxl_image.c36 unsigned int chunk_size) in qxl_allocate_chunk() argument
45 ret = qxl_alloc_bo_reserved(qdev, release, chunk_size, &chunk->bo); in qxl_allocate_chunk()
/drivers/net/wireless/intel/iwlwifi/mvm/
Dnvm.c215 int chunk_size, ret; in iwl_nvm_write_section() local
217 chunk_size = min(IWL_NVM_DEFAULT_CHUNK_SIZE, in iwl_nvm_write_section()
221 chunk_size, data + offset); in iwl_nvm_write_section()
225 offset += chunk_size; in iwl_nvm_write_section()
/drivers/infiniband/hw/mthca/
Dmthca_memfree.c364 unsigned chunk_size; in mthca_alloc_icm_table() local
386 chunk_size = MTHCA_TABLE_CHUNK_SIZE; in mthca_alloc_icm_table()
388 chunk_size = nobj * obj_size - i * MTHCA_TABLE_CHUNK_SIZE; in mthca_alloc_icm_table()
390 table->icm[i] = mthca_alloc_icm(dev, chunk_size >> PAGE_SHIFT, in mthca_alloc_icm_table()
/drivers/nvme/host/
Dpci.c1642 u32 chunk_size) in __nvme_alloc_host_mem() argument
1651 tmp = (preferred + chunk_size - 1); in __nvme_alloc_host_mem()
1652 do_div(tmp, chunk_size); in __nvme_alloc_host_mem()
1670 len = min_t(u64, chunk_size, preferred - size); in __nvme_alloc_host_mem()
1711 u32 chunk_size; in nvme_alloc_host_mem() local
1714 for (chunk_size = min_t(u64, preferred, PAGE_SIZE * MAX_ORDER_NR_PAGES); in nvme_alloc_host_mem()
1715 chunk_size >= max_t(u32, dev->ctrl.hmminds * 4096, PAGE_SIZE * 2); in nvme_alloc_host_mem()
1716 chunk_size /= 2) { in nvme_alloc_host_mem()
1717 if (!__nvme_alloc_host_mem(dev, preferred, chunk_size)) { in nvme_alloc_host_mem()

12