Lines Matching refs:data
499 static int crc32_threadfn(void *data) in crc32_threadfn() argument
501 struct crc_data *d = data; in crc32_threadfn()
543 static int lzo_compress_threadfn(void *data) in lzo_compress_threadfn() argument
545 struct cmp_data *d = data; in lzo_compress_threadfn()
588 struct cmp_data *data = NULL; in save_image_lzo() local
605 data = vmalloc(sizeof(*data) * nr_threads); in save_image_lzo()
606 if (!data) { in save_image_lzo()
612 memset(&data[thr], 0, offsetof(struct cmp_data, go)); in save_image_lzo()
626 init_waitqueue_head(&data[thr].go); in save_image_lzo()
627 init_waitqueue_head(&data[thr].done); in save_image_lzo()
629 data[thr].thr = kthread_run(lzo_compress_threadfn, in save_image_lzo()
630 &data[thr], in save_image_lzo()
632 if (IS_ERR(data[thr].thr)) { in save_image_lzo()
633 data[thr].thr = NULL; in save_image_lzo()
650 crc->unc[thr] = data[thr].unc; in save_image_lzo()
651 crc->unc_len[thr] = &data[thr].unc_len; in save_image_lzo()
688 memcpy(data[thr].unc + off, in save_image_lzo()
701 data[thr].unc_len = off; in save_image_lzo()
703 atomic_set(&data[thr].ready, 1); in save_image_lzo()
704 wake_up(&data[thr].go); in save_image_lzo()
715 wait_event(data[thr].done, in save_image_lzo()
716 atomic_read(&data[thr].stop)); in save_image_lzo()
717 atomic_set(&data[thr].stop, 0); in save_image_lzo()
719 ret = data[thr].ret; in save_image_lzo()
726 if (unlikely(!data[thr].cmp_len || in save_image_lzo()
727 data[thr].cmp_len > in save_image_lzo()
728 lzo1x_worst_compress(data[thr].unc_len))) { in save_image_lzo()
735 *(size_t *)data[thr].cmp = data[thr].cmp_len; in save_image_lzo()
746 off < LZO_HEADER + data[thr].cmp_len; in save_image_lzo()
748 memcpy(page, data[thr].cmp + off, PAGE_SIZE); in save_image_lzo()
774 if (data) { in save_image_lzo()
776 if (data[thr].thr) in save_image_lzo()
777 kthread_stop(data[thr].thr); in save_image_lzo()
778 vfree(data); in save_image_lzo()
1031 static int lzo_decompress_threadfn(void *data) in lzo_decompress_threadfn() argument
1033 struct dec_data *d = data; in lzo_decompress_threadfn()
1079 struct dec_data *data = NULL; in load_image_lzo() local
1096 data = vmalloc(sizeof(*data) * nr_threads); in load_image_lzo()
1097 if (!data) { in load_image_lzo()
1103 memset(&data[thr], 0, offsetof(struct dec_data, go)); in load_image_lzo()
1117 init_waitqueue_head(&data[thr].go); in load_image_lzo()
1118 init_waitqueue_head(&data[thr].done); in load_image_lzo()
1120 data[thr].thr = kthread_run(lzo_decompress_threadfn, in load_image_lzo()
1121 &data[thr], in load_image_lzo()
1123 if (IS_ERR(data[thr].thr)) { in load_image_lzo()
1124 data[thr].thr = NULL; in load_image_lzo()
1141 crc->unc[thr] = data[thr].unc; in load_image_lzo()
1142 crc->unc_len[thr] = &data[thr].unc_len; in load_image_lzo()
1244 data[thr].cmp_len = *(size_t *)page[pg]; in load_image_lzo()
1245 if (unlikely(!data[thr].cmp_len || in load_image_lzo()
1246 data[thr].cmp_len > in load_image_lzo()
1254 need = DIV_ROUND_UP(data[thr].cmp_len + LZO_HEADER, in load_image_lzo()
1265 off < LZO_HEADER + data[thr].cmp_len; in load_image_lzo()
1267 memcpy(data[thr].cmp + off, in load_image_lzo()
1275 atomic_set(&data[thr].ready, 1); in load_image_lzo()
1276 wake_up(&data[thr].go); in load_image_lzo()
1293 wait_event(data[thr].done, in load_image_lzo()
1294 atomic_read(&data[thr].stop)); in load_image_lzo()
1295 atomic_set(&data[thr].stop, 0); in load_image_lzo()
1297 ret = data[thr].ret; in load_image_lzo()
1305 if (unlikely(!data[thr].unc_len || in load_image_lzo()
1306 data[thr].unc_len > LZO_UNC_SIZE || in load_image_lzo()
1307 data[thr].unc_len & (PAGE_SIZE - 1))) { in load_image_lzo()
1315 off < data[thr].unc_len; off += PAGE_SIZE) { in load_image_lzo()
1317 data[thr].unc + off, PAGE_SIZE); in load_image_lzo()
1371 if (data) { in load_image_lzo()
1373 if (data[thr].thr) in load_image_lzo()
1374 kthread_stop(data[thr].thr); in load_image_lzo()
1375 vfree(data); in load_image_lzo()