Lines Matching refs:thr
579 struct task_struct *thr; /* thread */ member
602 d->thr = NULL; in crc32_threadfn()
621 struct task_struct *thr; /* thread */ member
645 d->thr = NULL; in lzo_compress_threadfn()
680 unsigned thr, run_threads, nr_threads; in save_image_lzo() local
707 for (thr = 0; thr < nr_threads; thr++) in save_image_lzo()
708 memset(&data[thr], 0, offsetof(struct cmp_data, go)); in save_image_lzo()
721 for (thr = 0; thr < nr_threads; thr++) { in save_image_lzo()
722 init_waitqueue_head(&data[thr].go); in save_image_lzo()
723 init_waitqueue_head(&data[thr].done); in save_image_lzo()
725 data[thr].thr = kthread_run(lzo_compress_threadfn, in save_image_lzo()
726 &data[thr], in save_image_lzo()
727 "image_compress/%u", thr); in save_image_lzo()
728 if (IS_ERR(data[thr].thr)) { in save_image_lzo()
729 data[thr].thr = NULL; in save_image_lzo()
744 for (thr = 0; thr < nr_threads; thr++) { in save_image_lzo()
745 crc->unc[thr] = data[thr].unc; in save_image_lzo()
746 crc->unc_len[thr] = &data[thr].unc_len; in save_image_lzo()
749 crc->thr = kthread_run(crc32_threadfn, crc, "image_crc32"); in save_image_lzo()
750 if (IS_ERR(crc->thr)) { in save_image_lzo()
751 crc->thr = NULL; in save_image_lzo()
772 for (thr = 0; thr < nr_threads; thr++) { in save_image_lzo()
781 memcpy(data[thr].unc + off, in save_image_lzo()
792 data[thr].unc_len = off; in save_image_lzo()
794 atomic_set(&data[thr].ready, 1); in save_image_lzo()
795 wake_up(&data[thr].go); in save_image_lzo()
798 if (!thr) in save_image_lzo()
801 crc->run_threads = thr; in save_image_lzo()
805 for (run_threads = thr, thr = 0; thr < run_threads; thr++) { in save_image_lzo()
806 wait_event(data[thr].done, in save_image_lzo()
807 atomic_read(&data[thr].stop)); in save_image_lzo()
808 atomic_set(&data[thr].stop, 0); in save_image_lzo()
810 ret = data[thr].ret; in save_image_lzo()
817 if (unlikely(!data[thr].cmp_len || in save_image_lzo()
818 data[thr].cmp_len > in save_image_lzo()
819 lzo1x_worst_compress(data[thr].unc_len))) { in save_image_lzo()
825 *(size_t *)data[thr].cmp = data[thr].cmp_len; in save_image_lzo()
836 off < LZO_HEADER + data[thr].cmp_len; in save_image_lzo()
838 memcpy(page, data[thr].cmp + off, PAGE_SIZE); in save_image_lzo()
860 if (crc->thr) in save_image_lzo()
861 kthread_stop(crc->thr); in save_image_lzo()
865 for (thr = 0; thr < nr_threads; thr++) in save_image_lzo()
866 if (data[thr].thr) in save_image_lzo()
867 kthread_stop(data[thr].thr); in save_image_lzo()
1107 struct task_struct *thr; /* thread */ member
1130 d->thr = NULL; in lzo_decompress_threadfn()
1169 unsigned i, thr, run_threads, nr_threads; in load_image_lzo() local
1199 for (thr = 0; thr < nr_threads; thr++) in load_image_lzo()
1200 memset(&data[thr], 0, offsetof(struct dec_data, go)); in load_image_lzo()
1215 for (thr = 0; thr < nr_threads; thr++) { in load_image_lzo()
1216 init_waitqueue_head(&data[thr].go); in load_image_lzo()
1217 init_waitqueue_head(&data[thr].done); in load_image_lzo()
1219 data[thr].thr = kthread_run(lzo_decompress_threadfn, in load_image_lzo()
1220 &data[thr], in load_image_lzo()
1221 "image_decompress/%u", thr); in load_image_lzo()
1222 if (IS_ERR(data[thr].thr)) { in load_image_lzo()
1223 data[thr].thr = NULL; in load_image_lzo()
1238 for (thr = 0; thr < nr_threads; thr++) { in load_image_lzo()
1239 crc->unc[thr] = data[thr].unc; in load_image_lzo()
1240 crc->unc_len[thr] = &data[thr].unc_len; in load_image_lzo()
1243 crc->thr = kthread_run(crc32_threadfn, crc, "image_crc32"); in load_image_lzo()
1244 if (IS_ERR(crc->thr)) { in load_image_lzo()
1245 crc->thr = NULL; in load_image_lzo()
1338 for (thr = 0; have && thr < nr_threads; thr++) { in load_image_lzo()
1339 data[thr].cmp_len = *(size_t *)page[pg]; in load_image_lzo()
1340 if (unlikely(!data[thr].cmp_len || in load_image_lzo()
1341 data[thr].cmp_len > in load_image_lzo()
1348 need = DIV_ROUND_UP(data[thr].cmp_len + LZO_HEADER, in load_image_lzo()
1359 off < LZO_HEADER + data[thr].cmp_len; in load_image_lzo()
1361 memcpy(data[thr].cmp + off, in load_image_lzo()
1369 atomic_set(&data[thr].ready, 1); in load_image_lzo()
1370 wake_up(&data[thr].go); in load_image_lzo()
1386 for (run_threads = thr, thr = 0; thr < run_threads; thr++) { in load_image_lzo()
1387 wait_event(data[thr].done, in load_image_lzo()
1388 atomic_read(&data[thr].stop)); in load_image_lzo()
1389 atomic_set(&data[thr].stop, 0); in load_image_lzo()
1391 ret = data[thr].ret; in load_image_lzo()
1398 if (unlikely(!data[thr].unc_len || in load_image_lzo()
1399 data[thr].unc_len > LZO_UNC_SIZE || in load_image_lzo()
1400 data[thr].unc_len & (PAGE_SIZE - 1))) { in load_image_lzo()
1407 off < data[thr].unc_len; off += PAGE_SIZE) { in load_image_lzo()
1409 data[thr].unc + off, PAGE_SIZE); in load_image_lzo()
1418 crc->run_threads = thr + 1; in load_image_lzo()
1426 crc->run_threads = thr; in load_image_lzo()
1456 if (crc->thr) in load_image_lzo()
1457 kthread_stop(crc->thr); in load_image_lzo()
1461 for (thr = 0; thr < nr_threads; thr++) in load_image_lzo()
1462 if (data[thr].thr) in load_image_lzo()
1463 kthread_stop(data[thr].thr); in load_image_lzo()