Lines Matching refs:tsz
225 size_t tsz; in vmcoredd_copy_dumps() local
231 tsz = min(offset + (u64)dump->size - start, (u64)size); in vmcoredd_copy_dumps()
233 if (copy_to(dst, buf, tsz, userbuf)) { in vmcoredd_copy_dumps()
238 size -= tsz; in vmcoredd_copy_dumps()
239 start += tsz; in vmcoredd_copy_dumps()
240 dst += tsz; in vmcoredd_copy_dumps()
261 size_t tsz; in vmcoredd_mmap_dumps() local
267 tsz = min(offset + (u64)dump->size - start, (u64)size); in vmcoredd_mmap_dumps()
269 if (remap_vmalloc_range_partial(vma, dst, buf, tsz)) { in vmcoredd_mmap_dumps()
274 size -= tsz; in vmcoredd_mmap_dumps()
275 start += tsz; in vmcoredd_mmap_dumps()
276 dst += tsz; in vmcoredd_mmap_dumps()
299 size_t tsz; in __read_vmcore() local
312 tsz = min(elfcorebuf_sz - (size_t)*fpos, buflen); in __read_vmcore()
313 if (copy_to(buffer, elfcorebuf + *fpos, tsz, userbuf)) in __read_vmcore()
315 buflen -= tsz; in __read_vmcore()
316 *fpos += tsz; in __read_vmcore()
317 buffer += tsz; in __read_vmcore()
318 acc += tsz; in __read_vmcore()
341 tsz = min(elfcorebuf_sz + vmcoredd_orig_sz - in __read_vmcore()
344 if (vmcoredd_copy_dumps(buffer, start, tsz, userbuf)) in __read_vmcore()
347 buflen -= tsz; in __read_vmcore()
348 *fpos += tsz; in __read_vmcore()
349 buffer += tsz; in __read_vmcore()
350 acc += tsz; in __read_vmcore()
359 tsz = min(elfcorebuf_sz + elfnotes_sz - (size_t)*fpos, buflen); in __read_vmcore()
361 if (copy_to(buffer, kaddr, tsz, userbuf)) in __read_vmcore()
364 buflen -= tsz; in __read_vmcore()
365 *fpos += tsz; in __read_vmcore()
366 buffer += tsz; in __read_vmcore()
367 acc += tsz; in __read_vmcore()
376 tsz = (size_t)min_t(unsigned long long, in __read_vmcore()
380 tmp = read_from_oldmem(buffer, tsz, &start, in __read_vmcore()
384 buflen -= tsz; in __read_vmcore()
385 *fpos += tsz; in __read_vmcore()
386 buffer += tsz; in __read_vmcore()
387 acc += tsz; in __read_vmcore()
553 u64 start, end, len, tsz; in mmap_vmcore() local
574 tsz = min(elfcorebuf_sz - (size_t)start, size); in mmap_vmcore()
576 if (remap_pfn_range(vma, vma->vm_start, pfn, tsz, in mmap_vmcore()
579 size -= tsz; in mmap_vmcore()
580 start += tsz; in mmap_vmcore()
581 len += tsz; in mmap_vmcore()
606 tsz = min(elfcorebuf_sz + vmcoredd_orig_sz - in mmap_vmcore()
610 start_off, tsz)) in mmap_vmcore()
613 size -= tsz; in mmap_vmcore()
614 start += tsz; in mmap_vmcore()
615 len += tsz; in mmap_vmcore()
624 tsz = min(elfcorebuf_sz + elfnotes_sz - (size_t)start, size); in mmap_vmcore()
627 kaddr, tsz)) in mmap_vmcore()
630 size -= tsz; in mmap_vmcore()
631 start += tsz; in mmap_vmcore()
632 len += tsz; in mmap_vmcore()
642 tsz = (size_t)min_t(unsigned long long, in mmap_vmcore()
646 paddr >> PAGE_SHIFT, tsz, in mmap_vmcore()
649 size -= tsz; in mmap_vmcore()
650 start += tsz; in mmap_vmcore()
651 len += tsz; in mmap_vmcore()