/drivers/md/ |
D | dm-exception-store.c | 145 unsigned chunk_size; in set_chunk_size() local 147 if (kstrtouint(chunk_size_arg, 10, &chunk_size)) { in set_chunk_size() 152 if (!chunk_size) { in set_chunk_size() 153 store->chunk_size = store->chunk_mask = store->chunk_shift = 0; in set_chunk_size() 157 return dm_exception_store_set_chunk_size(store, chunk_size, error); in set_chunk_size() 161 unsigned chunk_size, in dm_exception_store_set_chunk_size() argument 165 if (!is_power_of_2(chunk_size)) { in dm_exception_store_set_chunk_size() 171 if (chunk_size % in dm_exception_store_set_chunk_size() 173 chunk_size % in dm_exception_store_set_chunk_size() 179 if (chunk_size > INT_MAX >> SECTOR_SHIFT) { in dm_exception_store_set_chunk_size() [all …]
|
D | dm-stripe.c | 35 uint32_t chunk_size; member 103 uint32_t chunk_size; in stripe_ctr() local 117 if (kstrtouint(argv[1], 10, &chunk_size) || !chunk_size) { in stripe_ctr() 130 if (sector_div(tmp_len, chunk_size)) { in stripe_ctr() 164 r = dm_set_target_max_io_len(ti, chunk_size); in stripe_ctr() 175 sc->chunk_size = chunk_size; in stripe_ctr() 176 if (chunk_size & (chunk_size - 1)) in stripe_ctr() 179 sc->chunk_size_shift = __ffs(chunk_size); in stripe_ctr() 222 chunk_offset = sector_div(chunk, sc->chunk_size); in stripe_map_sector() 224 chunk_offset = chunk & (sc->chunk_size - 1); in stripe_map_sector() [all …]
|
D | dm-snap-persistent.c | 81 __le32 chunk_size; member 172 len = ps->store->chunk_size << SECTOR_SHIFT; in alloc_area() 234 .sector = ps->store->chunk_size * chunk, in chunk_io() 235 .count = ps->store->chunk_size, in chunk_io() 301 memset(ps->area, 0, ps->store->chunk_size << SECTOR_SHIFT); in zero_memory_area() 314 unsigned chunk_size; in read_header() local 322 if (!ps->store->chunk_size) { in read_header() 323 ps->store->chunk_size = max(DM_CHUNK_SIZE_DEFAULT_SECTORS, in read_header() 326 ps->store->chunk_mask = ps->store->chunk_size - 1; in read_header() 327 ps->store->chunk_shift = __ffs(ps->store->chunk_size); in read_header() [all …]
|
D | dm-snap-transient.c | 45 if (size < (tc->next_free + store->chunk_size)) in transient_prepare_exception() 49 tc->next_free += store->chunk_size; in transient_prepare_exception() 97 DMEMIT(" N %llu", (unsigned long long)store->chunk_size); in transient_status()
|
D | dm-exception-store.h | 120 unsigned chunk_size; member 207 unsigned chunk_size,
|
D | dm-snap.c | 530 if (l->store->chunk_size < s->store->chunk_size) in __insert_snapshot() 791 unsigned chunk_size = 0; in __minimum_chunk_size() local 795 chunk_size = min_not_zero(chunk_size, in __minimum_chunk_size() 796 snap->store->chunk_size); in __minimum_chunk_size() 798 return (uint32_t) chunk_size; in __minimum_chunk_size() 951 sector_t sector, unsigned chunk_size); 1017 io_size = linear_chunks * s->store->chunk_size; in snapshot_merge_next_chunks() 1251 s->store->chunk_size = 0; in snapshot_ctr() 1265 if (!s->store->chunk_size) { in snapshot_ctr() 1270 r = dm_set_target_max_io_len(ti, s->store->chunk_size); in snapshot_ctr() [all …]
|
/drivers/staging/lustre/lustre/obdclass/ |
D | llog.c | 149 int chunk_size = handle->lgh_ctxt->loc_chunk_size; in llog_init_handle() local 156 LASSERT(chunk_size >= LLOG_MIN_CHUNK_SIZE); in llog_init_handle() 161 handle->lgh_hdr_size = chunk_size; in llog_init_handle() 229 int chunk_size; in llog_process_thread() local 238 chunk_size = llh->llh_hdr.lrh_len; in llog_process_thread() 240 LASSERT(is_power_of_2(chunk_size)); in llog_process_thread() 242 buf = libcfs_kvzalloc(chunk_size, GFP_NOFS); in llog_process_thread() 275 memset(buf, 0, chunk_size); in llog_process_thread() 277 index, &cur_offset, buf, chunk_size); in llog_process_thread() 288 if (do_div(tmp_offset, chunk_size)) { in llog_process_thread() [all …]
|
/drivers/staging/android/ion/ |
D | ion_chunk_heap.c | 30 unsigned long chunk_size; member 48 allocated_size = ALIGN(size, chunk_heap->chunk_size); in ion_chunk_heap_allocate() 49 num_chunks = allocated_size / chunk_heap->chunk_size; in ion_chunk_heap_allocate() 66 chunk_heap->chunk_size); in ion_chunk_heap_allocate() 70 chunk_heap->chunk_size, 0); in ion_chunk_heap_allocate() 99 allocated_size = ALIGN(buffer->size, chunk_heap->chunk_size); in ion_chunk_heap_free() 138 chunk_heap->chunk_size = (unsigned long)heap_data->priv; in ion_chunk_heap_create() 139 chunk_heap->pool = gen_pool_create(get_order(chunk_heap->chunk_size) + in ion_chunk_heap_create()
|
/drivers/net/ethernet/mellanox/mlxsw/ |
D | i2c.c | 285 int off = mlxsw_i2c->cmd.mb_off_in, chunk_size, i, j; in mlxsw_i2c_write() local 292 chunk_size = (in_mbox_size > MLXSW_I2C_BLK_MAX) ? in mlxsw_i2c_write() 294 write_tran.len = MLXSW_I2C_ADDR_WIDTH + chunk_size; in mlxsw_i2c_write() 297 MLXSW_I2C_BLK_MAX * i, chunk_size); in mlxsw_i2c_write() 316 off += chunk_size; in mlxsw_i2c_write() 317 in_mbox_size -= chunk_size; in mlxsw_i2c_write() 353 int num, chunk_size, reg_size, i, j; in mlxsw_i2c_cmd() local 385 chunk_size = (reg_size > MLXSW_I2C_BLK_MAX) ? in mlxsw_i2c_cmd() 387 read_tran[1].len = chunk_size; in mlxsw_i2c_cmd() 409 off += chunk_size; in mlxsw_i2c_cmd() [all …]
|
/drivers/media/rc/ |
D | lirc_dev.c | 45 unsigned int chunk_size; member 79 unsigned int chunk_size; in lirc_allocate_buffer() local 86 chunk_size = d->chunk_size ? d->chunk_size : bytes_in_key; in lirc_allocate_buffer() 98 err = lirc_buffer_init(ir->buf, chunk_size, buffer_size); in lirc_allocate_buffer() 108 ir->chunk_size = ir->buf->chunk_size; in lirc_allocate_buffer() 491 buf = kzalloc(ir->chunk_size, GFP_KERNEL); in lirc_dev_fop_read() 504 if (length % ir->chunk_size) { in lirc_dev_fop_read() 556 ir->buf->chunk_size); in lirc_dev_fop_read() 558 written += ir->buf->chunk_size; in lirc_dev_fop_read()
|
/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_device.c | 92 unsigned int chunk_size); 403 unsigned int chunk_size) in kfd_gtt_sa_init() argument 407 if (WARN_ON(buf_size < chunk_size)) in kfd_gtt_sa_init() 411 if (WARN_ON(chunk_size == 0)) in kfd_gtt_sa_init() 414 kfd->gtt_sa_chunk_size = chunk_size; in kfd_gtt_sa_init() 415 kfd->gtt_sa_num_of_chunks = buf_size / chunk_size; in kfd_gtt_sa_init() 442 unsigned int chunk_size) in kfd_gtt_sa_calc_gpu_addr() argument 444 return start_addr + bit_num * chunk_size; in kfd_gtt_sa_calc_gpu_addr() 449 unsigned int chunk_size) in kfd_gtt_sa_calc_cpu_addr() argument 451 return (uint32_t *) ((uint64_t) start_addr + bit_num * chunk_size); in kfd_gtt_sa_calc_cpu_addr()
|
/drivers/net/wireless/marvell/libertas/ |
D | if_sdio.c | 448 u32 chunk_size; in if_sdio_prog_helper() local 477 chunk_size = min_t(size_t, size, 60); in if_sdio_prog_helper() 479 *((__le32*)chunk_buffer) = cpu_to_le32(chunk_size); in if_sdio_prog_helper() 480 memcpy(chunk_buffer + 4, firmware, chunk_size); in if_sdio_prog_helper() 489 firmware += chunk_size; in if_sdio_prog_helper() 490 size -= chunk_size; in if_sdio_prog_helper() 544 u32 chunk_size; in if_sdio_prog_real() local 615 chunk_size = min_t(size_t, req_size, 512); in if_sdio_prog_real() 617 memcpy(chunk_buffer, firmware, chunk_size); in if_sdio_prog_real() 623 chunk_buffer, roundup(chunk_size, 32)); in if_sdio_prog_real() [all …]
|
/drivers/net/wireless/ath/wcn36xx/ |
D | dxe.c | 242 int i, chunk_size = pool->chunk_size; in wcn36xx_dxe_init_tx_bd() local 253 bd_phy_addr += chunk_size; in wcn36xx_dxe_init_tx_bd() 254 bd_cpu_addr += chunk_size; in wcn36xx_dxe_init_tx_bd() 554 wcn->mgmt_mem_pool.chunk_size = WCN36XX_BD_CHUNK_SIZE + in wcn36xx_dxe_allocate_mem_pools() 557 s = wcn->mgmt_mem_pool.chunk_size * WCN36XX_DXE_CH_DESC_NUMB_TX_H; in wcn36xx_dxe_allocate_mem_pools() 569 wcn->data_mem_pool.chunk_size = WCN36XX_BD_CHUNK_SIZE + in wcn36xx_dxe_allocate_mem_pools() 572 s = wcn->data_mem_pool.chunk_size * WCN36XX_DXE_CH_DESC_NUMB_TX_L; in wcn36xx_dxe_allocate_mem_pools() 592 dma_free_coherent(wcn->dev, wcn->mgmt_mem_pool.chunk_size * in wcn36xx_dxe_free_mem_pools() 598 dma_free_coherent(wcn->dev, wcn->data_mem_pool.chunk_size * in wcn36xx_dxe_free_mem_pools()
|
/drivers/base/regmap/ |
D | regmap.c | 1914 size_t chunk_size = val_bytes; in regmap_bulk_write() local 1918 chunk_size = map->max_raw_write; in regmap_bulk_write() 1919 if (chunk_size % val_bytes) in regmap_bulk_write() 1920 chunk_size -= chunk_size % val_bytes; in regmap_bulk_write() 1921 chunk_count = total_size / chunk_size; in regmap_bulk_write() 1922 chunk_stride *= chunk_size / val_bytes; in regmap_bulk_write() 1930 val + (i * chunk_size), in regmap_bulk_write() 1931 chunk_size); in regmap_bulk_write() 1937 if (!ret && chunk_size * i < total_size) { in regmap_bulk_write() 1939 val + (i * chunk_size), in regmap_bulk_write() [all …]
|
/drivers/rpmsg/ |
D | qcom_glink_native.c | 790 __le32 chunk_size; in qcom_glink_rx_data() member 793 unsigned int chunk_size; in qcom_glink_rx_data() local 806 chunk_size = le32_to_cpu(hdr.chunk_size); in qcom_glink_rx_data() 809 if (avail < sizeof(hdr) + chunk_size) { in qcom_glink_rx_data() 814 if (WARN(chunk_size % 4, "Incoming data must be word aligned\n")) in qcom_glink_rx_data() 835 intent->data = kmalloc(chunk_size + left_size, in qcom_glink_rx_data() 843 intent->size = chunk_size + left_size; in qcom_glink_rx_data() 865 if (intent->size - intent->offset < chunk_size) { in qcom_glink_rx_data() 873 sizeof(hdr), chunk_size); in qcom_glink_rx_data() 874 intent->offset += chunk_size; in qcom_glink_rx_data() [all …]
|
/drivers/mtd/nand/ |
D | pxa3xx_nand.c | 236 unsigned int chunk_size; member 332 int nchunks = mtd->writesize / info->chunk_size; in pxa3xx_ooblayout_ecc() 350 int nchunks = mtd->writesize / info->chunk_size; in pxa3xx_ooblayout_free() 365 if (mtd->writesize == 4096 && info->chunk_size == 2048) { in pxa3xx_ooblayout_free() 776 info->sg.length = info->chunk_size; in start_data_dma() 1015 info->step_chunk_size = info->chunk_size; in prepare_set_command() 1066 info->step_chunk_size = info->chunk_size; in prepare_set_command() 1465 info->chunk_size = PAGE_CHUNK_SIZE; in pxa3xx_nand_config_ident() 1498 info->chunk_size = ndcr & NDCR_PAGE_SZ ? 2048 : 512; in pxa3xx_nand_detect_config() 1579 info->chunk_size = 2048; in pxa_ecc_init() [all …]
|
/drivers/gpu/drm/vmwgfx/ |
D | vmwgfx_fifo.c | 413 uint32_t chunk_size = max - next_cmd; in vmw_fifo_res_copy() local 418 if (bytes < chunk_size) in vmw_fifo_res_copy() 419 chunk_size = bytes; in vmw_fifo_res_copy() 423 memcpy(fifo_mem + (next_cmd >> 2), buffer, chunk_size); in vmw_fifo_res_copy() 424 rest = bytes - chunk_size; in vmw_fifo_res_copy() 426 memcpy(fifo_mem + (min >> 2), buffer + (chunk_size >> 2), rest); in vmw_fifo_res_copy()
|
/drivers/net/ethernet/mellanox/mlx4/ |
D | icm.c | 396 unsigned chunk_size; in mlx4_init_icm_table() local 416 chunk_size = MLX4_TABLE_CHUNK_SIZE; in mlx4_init_icm_table() 418 chunk_size = PAGE_ALIGN(size - in mlx4_init_icm_table() 421 table->icm[i] = mlx4_alloc_icm(dev, chunk_size >> PAGE_SHIFT, in mlx4_init_icm_table()
|
/drivers/staging/media/lirc/ |
D | lirc_zilog.c | 893 if (n % rbuf->chunk_size) { in read() 937 if (rbuf->chunk_size > sizeof(buf)) { in read() 940 rbuf->chunk_size); in read() 945 if (m == rbuf->chunk_size) { in read() 947 rbuf->chunk_size); in read() 948 written += rbuf->chunk_size; in read() 1392 .chunk_size = 2, 1498 ir->l.chunk_size, ir->l.buffer_size); in ir_probe()
|
/drivers/uwb/i1480/dfu/ |
D | mac.c | 172 size_t chunk_size; in i1480_fw_cmp() local 176 chunk_size = size < i1480->buf_size ? size : i1480->buf_size; in i1480_fw_cmp() 177 result = i1480->read(i1480, hdr->address + src_itr, chunk_size); in i1480_fw_cmp()
|
/drivers/media/dvb-frontends/ |
D | drxk.h | 48 int chunk_size; member
|
/drivers/gpu/drm/qxl/ |
D | qxl_image.c | 36 unsigned int chunk_size) in qxl_allocate_chunk() argument 45 ret = qxl_alloc_bo_reserved(qdev, release, chunk_size, &chunk->bo); in qxl_allocate_chunk()
|
/drivers/net/wireless/intel/iwlwifi/mvm/ |
D | nvm.c | 215 int chunk_size, ret; in iwl_nvm_write_section() local 217 chunk_size = min(IWL_NVM_DEFAULT_CHUNK_SIZE, in iwl_nvm_write_section() 221 chunk_size, data + offset); in iwl_nvm_write_section() 225 offset += chunk_size; in iwl_nvm_write_section()
|
/drivers/infiniband/hw/mthca/ |
D | mthca_memfree.c | 364 unsigned chunk_size; in mthca_alloc_icm_table() local 386 chunk_size = MTHCA_TABLE_CHUNK_SIZE; in mthca_alloc_icm_table() 388 chunk_size = nobj * obj_size - i * MTHCA_TABLE_CHUNK_SIZE; in mthca_alloc_icm_table() 390 table->icm[i] = mthca_alloc_icm(dev, chunk_size >> PAGE_SHIFT, in mthca_alloc_icm_table()
|
/drivers/nvme/host/ |
D | pci.c | 1642 u32 chunk_size) in __nvme_alloc_host_mem() argument 1651 tmp = (preferred + chunk_size - 1); in __nvme_alloc_host_mem() 1652 do_div(tmp, chunk_size); in __nvme_alloc_host_mem() 1670 len = min_t(u64, chunk_size, preferred - size); in __nvme_alloc_host_mem() 1711 u32 chunk_size; in nvme_alloc_host_mem() local 1714 for (chunk_size = min_t(u64, preferred, PAGE_SIZE * MAX_ORDER_NR_PAGES); in nvme_alloc_host_mem() 1715 chunk_size >= max_t(u32, dev->ctrl.hmminds * 4096, PAGE_SIZE * 2); in nvme_alloc_host_mem() 1716 chunk_size /= 2) { in nvme_alloc_host_mem() 1717 if (!__nvme_alloc_host_mem(dev, preferred, chunk_size)) { in nvme_alloc_host_mem()
|