Lines Matching refs:r_addend

187 	*link_addr_out = (intptr_t)(link_addr + candidate.r_addend - candidate_loc); /* S + A - P */  in llext_riscv_find_sym_pcrel()
251 __typeof__(rel->r_addend) target_alignment = 1; in arch_elf_relocate()
279 jump_target = sym_base_addr + rel->r_addend; /* S + A */ in arch_elf_relocate()
284 UNALIGNED_PUT(sym_base_addr + rel->r_addend, loc64); /* S + A */ in arch_elf_relocate()
288 UNALIGNED_PUT(load_bias + rel->r_addend, loc_word); /* B + A */ in arch_elf_relocate()
295 jump_target = sym_base_addr + rel->r_addend - loc; /* S + A - P */ in arch_elf_relocate()
304 jump_target = sym_base_addr + rel->r_addend - loc; /* S + A - P */ in arch_elf_relocate()
316 jump_target = sym_base_addr + rel->r_addend - loc; /* S + A - P */ in arch_elf_relocate()
371 jump_target = sym_base_addr + rel->r_addend; /* S + A */ in arch_elf_relocate()
388 jump_target = sym_base_addr + rel->r_addend; /* S + A */ in arch_elf_relocate()
398 imm8 = sym_base_addr + rel->r_addend; /* S + A */ in arch_elf_relocate()
411 tmp8 += sym_base_addr + rel->r_addend; /* V + S + A */ in arch_elf_relocate()
416 tmp16 += sym_base_addr + rel->r_addend; /* V + S + A */ in arch_elf_relocate()
421 tmp32 += sym_base_addr + rel->r_addend; /* V + S + A */ in arch_elf_relocate()
426 tmp64 += sym_base_addr + rel->r_addend; /* V + S + A */ in arch_elf_relocate()
431 tmp8 -= sym_base_addr + rel->r_addend; /* V - S - A */ in arch_elf_relocate()
436 tmp16 -= sym_base_addr + rel->r_addend; /* V - S - A */ in arch_elf_relocate()
441 tmp32 -= sym_base_addr + rel->r_addend; /* V - S - A */ in arch_elf_relocate()
446 tmp64 -= sym_base_addr + rel->r_addend; /* V - S - A */ in arch_elf_relocate()
452 tmp8 = tmp8 - sym_base_addr - rel->r_addend; /* V - S - A */ in arch_elf_relocate()
460 tmp8 = sym_base_addr + rel->r_addend; /* S + A */ in arch_elf_relocate()
465 tmp8 = sym_base_addr + rel->r_addend; /* S + A */ in arch_elf_relocate()
469 tmp16 = sym_base_addr + rel->r_addend; /* S + A */ in arch_elf_relocate()
473 tmp32 = sym_base_addr + rel->r_addend; /* S + A */ in arch_elf_relocate()
477 jump_target = sym_base_addr + rel->r_addend - loc; /* S + A - P */ in arch_elf_relocate()
483 jump_target = sym_base_addr + rel->r_addend - loc; /* S + A - P */ in arch_elf_relocate()
489 jump_target = sym_base_addr + rel->r_addend - loc; /* S + A - P */ in arch_elf_relocate()
500 jump_target = sym_base_addr + rel->r_addend - loc; /* S + A - P */ in arch_elf_relocate()
515 while (target_alignment < rel->r_addend) { in arch_elf_relocate()