Lines Matching refs:asize
2674 u32 asize = le32_to_cpu(attr->size); in check_attr() local
2680 if (asize >= sbi->record_size || in check_attr()
2681 asize + PtrOffset(rec, attr) >= sbi->record_size || in check_attr()
2684 asize)) { in check_attr()
2692 if (rsize >= asize || in check_attr()
2693 le16_to_cpu(attr->res.data_off) + rsize > asize) { in check_attr()
2704 if (svcn > evcn + 1 || run_off >= asize || in check_attr()
2710 if (run_off > asize) in check_attr()
2714 Add2Ptr(attr, run_off), asize - run_off) < 0) { in check_attr()
2727 attr, le16_to_cpu(attr->res.data_off))) > asize) { in check_attr()
2821 u32 asize; in check_if_attr() local
2826 asize = le32_to_cpu(attr->size); in check_if_attr()
2827 if (!asize) in check_if_attr()
2830 o += asize; in check_if_attr()
2831 attr = Add2Ptr(attr, asize); in check_if_attr()
2845 u32 asize; in check_if_index_root() local
2850 asize = le32_to_cpu(attr->size); in check_if_index_root()
2851 if (!asize) in check_if_index_root()
2854 o += asize; in check_if_index_root()
2855 attr = Add2Ptr(attr, asize); in check_if_index_root()
2869 u32 asize = le32_to_cpu(attr->size); in check_if_root_index() local
2874 if (o >= asize) in check_if_root_index()
2917 u32 asize = le32_to_cpu(attr->size); in change_attr_size() local
2918 int dsize = nsize - asize; in change_attr_size()
2919 u8 *next = Add2Ptr(attr, asize); in change_attr_size()
2975 u32 asize = name_size + in attr_create_nonres_log() local
2978 attr = kzalloc(asize, GFP_NOFS); in attr_create_nonres_log()
2983 attr->size = cpu_to_le32(asize); in attr_create_nonres_log()
3034 u32 nsize, t32, asize, used, esize, off, bits; in do_action() local
3215 asize = le32_to_cpu(attr2->size); in do_action()
3219 !IS_ALIGNED(asize, 8) || in do_action()
3220 Add2Ptr(attr2, asize) > Add2Ptr(lrh, rec_len) || in do_action()
3225 memmove(Add2Ptr(attr, asize), attr, used - roff); in do_action()
3226 memcpy(attr, attr2, asize); in do_action()
3228 rec->used = cpu_to_le32(used + asize); in do_action()
3251 asize = le32_to_cpu(attr->size); in do_action()
3257 rec->used = cpu_to_le32(used - asize); in do_action()
3261 memmove(attr, Add2Ptr(attr, asize), used - asize - roff); in do_action()
3272 asize = le32_to_cpu(attr->size); in do_action()
3276 if (nsize > asize) in do_action()
3281 if (nsize > asize && nsize - asize > record_size - used) in do_action()
3287 if (nsize < asize) { in do_action()
3292 memmove(Add2Ptr(attr, nsize), Add2Ptr(attr, asize), in do_action()
3293 used - le16_to_cpu(lrh->record_off) - asize); in do_action()
3295 rec->used = cpu_to_le32(used + nsize - asize); in do_action()
3320 asize = le32_to_cpu(attr->size); in do_action()
3324 aoff < le16_to_cpu(attr->nres.run_off) || aoff > asize || in do_action()
3325 (nsize > asize && nsize - asize > record_size - used)) { in do_action()
3331 memmove(Add2Ptr(attr, nsize), Add2Ptr(attr, asize), in do_action()
3332 used - le16_to_cpu(lrh->record_off) - asize); in do_action()
3333 rec->used = cpu_to_le32(used + nsize - asize); in do_action()