Lines Matching refs:buff
202 kprobe_opcode_t *op_callback_addr, *emulate_step_addr, *buff; in arch_prepare_optimized_kprobe() local
214 buff = get_ppc_optinsn_slot(); in arch_prepare_optimized_kprobe()
215 if (!buff) in arch_prepare_optimized_kprobe()
227 b_offset = (unsigned long)buff - (unsigned long)p->addr; in arch_prepare_optimized_kprobe()
232 b_offset = (unsigned long)(buff + TMPL_RET_IDX) - in arch_prepare_optimized_kprobe()
240 pr_devel("Copying template to %p, size %lu\n", buff, size); in arch_prepare_optimized_kprobe()
242 rc = patch_instruction((struct ppc_inst *)(buff + i), in arch_prepare_optimized_kprobe()
252 patch_imm64_load_insns((unsigned long)op, 3, buff + TMPL_OP_IDX); in arch_prepare_optimized_kprobe()
265 (struct ppc_inst *)(buff + TMPL_CALL_HDLR_IDX), in arch_prepare_optimized_kprobe()
270 (struct ppc_inst *)(buff + TMPL_EMULATE_IDX), in arch_prepare_optimized_kprobe()
277 patch_instruction((struct ppc_inst *)(buff + TMPL_CALL_HDLR_IDX), in arch_prepare_optimized_kprobe()
279 patch_instruction((struct ppc_inst *)(buff + TMPL_EMULATE_IDX), in arch_prepare_optimized_kprobe()
286 patch_imm64_load_insns(ppc_inst_as_u64(temp), 4, buff + TMPL_INSN_IDX); in arch_prepare_optimized_kprobe()
291 patch_branch((struct ppc_inst *)(buff + TMPL_RET_IDX), (unsigned long)nip, 0); in arch_prepare_optimized_kprobe()
293 flush_icache_range((unsigned long)buff, in arch_prepare_optimized_kprobe()
294 (unsigned long)(&buff[TMPL_END_IDX])); in arch_prepare_optimized_kprobe()
296 op->optinsn.insn = buff; in arch_prepare_optimized_kprobe()
301 free_ppc_optinsn_slot(buff, 0); in arch_prepare_optimized_kprobe()