Lines Matching refs:pr_err
103 pr_err("%s: unknown immediate encoding %d\n", __func__, in aarch64_insn_decode_immediate()
133 pr_err("%s: unknown immediate encoding %d\n", __func__, in aarch64_insn_encode_immediate()
167 pr_err("%s: unknown register type encoding %d\n", __func__, in aarch64_insn_decode_register()
185 pr_err("%s: unknown register encoding %d\n", __func__, reg); in aarch64_insn_encode_register()
206 pr_err("%s: unknown register type encoding %d\n", __func__, in aarch64_insn_encode_register()
230 pr_err("%s: unknown size encoding %d\n", __func__, type); in aarch64_insn_encode_ldst_size()
247 pr_err("%s: A64 instructions must be word aligned\n", __func__); in label_imm_common()
254 pr_err("%s: offset out of range\n", __func__); in label_imm_common()
284 pr_err("%s: unknown branch encoding %d\n", __func__, type); in aarch64_insn_gen_branch_imm()
312 pr_err("%s: unknown branch encoding %d\n", __func__, type); in aarch64_insn_gen_comp_branch_imm()
323 pr_err("%s: unknown variant encoding %d\n", __func__, variant); in aarch64_insn_gen_comp_branch_imm()
344 pr_err("%s: unknown condition encoding %d\n", __func__, cond); in aarch64_insn_gen_cond_branch_imm()
369 pr_err("%s: unknown branch encoding %d\n", __func__, type); in aarch64_insn_gen_branch_reg()
395 pr_err("%s: unknown load/store encoding %d\n", __func__, type); in aarch64_insn_gen_load_store_reg()
420 pr_err("%s: unknown size encoding %d\n", __func__, type); in aarch64_insn_gen_load_store_imm()
426 pr_err("%s: invalid imm: %d\n", __func__, imm); in aarch64_insn_gen_load_store_imm()
443 pr_err("%s: unknown load/store encoding %d\n", __func__, type); in aarch64_insn_gen_load_store_imm()
503 pr_err("%s: unknown load/store encoding %d\n", __func__, type); in aarch64_insn_gen_load_store_pair()
510 pr_err("%s: offset must be multiples of 4 in the range of [-256, 252] %d\n", in aarch64_insn_gen_load_store_pair()
518 pr_err("%s: offset must be multiples of 8 in the range of [-512, 504] %d\n", in aarch64_insn_gen_load_store_pair()
526 pr_err("%s: unknown variant encoding %d\n", __func__, variant); in aarch64_insn_gen_load_store_pair()
565 pr_err("%s: unknown load/store exclusive encoding %d\n", __func__, type); in aarch64_insn_gen_load_store_ex()
604 pr_err("%s: unknown mem order %d\n", __func__, type); in aarch64_insn_encode_ldst_order()
640 pr_err("%s: unimplemented mem atomic op %d\n", __func__, op); in aarch64_insn_gen_atomic_ld_op()
649 pr_err("%s: unimplemented size encoding %d\n", __func__, size); in aarch64_insn_gen_atomic_ld_op()
686 pr_err("%s: unknown mem order %d\n", __func__, type); in aarch64_insn_encode_cas_order()
709 pr_err("%s: unimplemented size encoding %d\n", __func__, size); in aarch64_insn_gen_cas()
751 pr_err("%s: unknown add/sub encoding %d\n", __func__, type); in aarch64_insn_gen_add_sub_imm()
762 pr_err("%s: unknown variant encoding %d\n", __func__, variant); in aarch64_insn_gen_add_sub_imm()
787 pr_err("%s: invalid immediate encoding %d\n", __func__, imm); in aarch64_insn_gen_add_sub_imm()
811 pr_err("%s: unknown bitfield encoding %d\n", __func__, type); in aarch64_insn_gen_bitfield()
824 pr_err("%s: unknown variant encoding %d\n", __func__, variant); in aarch64_insn_gen_bitfield()
829 pr_err("%s: invalid immr encoding %d\n", __func__, immr); in aarch64_insn_gen_bitfield()
833 pr_err("%s: invalid imms encoding %d\n", __func__, imms); in aarch64_insn_gen_bitfield()
864 pr_err("%s: unknown movewide encoding %d\n", __func__, type); in aarch64_insn_gen_movewide()
869 pr_err("%s: invalid immediate encoding %d\n", __func__, imm); in aarch64_insn_gen_movewide()
876 pr_err("%s: invalid shift encoding %d\n", __func__, in aarch64_insn_gen_movewide()
884 pr_err("%s: invalid shift encoding %d\n", __func__, in aarch64_insn_gen_movewide()
890 pr_err("%s: unknown variant encoding %d\n", __func__, variant); in aarch64_insn_gen_movewide()
924 pr_err("%s: unknown add/sub encoding %d\n", __func__, type); in aarch64_insn_gen_add_sub_shifted_reg()
931 pr_err("%s: invalid shift encoding %d\n", __func__, in aarch64_insn_gen_add_sub_shifted_reg()
939 pr_err("%s: invalid shift encoding %d\n", __func__, in aarch64_insn_gen_add_sub_shifted_reg()
945 pr_err("%s: unknown variant encoding %d\n", __func__, variant); in aarch64_insn_gen_add_sub_shifted_reg()
975 pr_err("%s: invalid variant for reverse64 %d\n", in aarch64_insn_gen_data1()
982 pr_err("%s: unknown data1 encoding %d\n", __func__, type); in aarch64_insn_gen_data1()
993 pr_err("%s: unknown variant encoding %d\n", __func__, variant); in aarch64_insn_gen_data1()
1030 pr_err("%s: unknown data2 encoding %d\n", __func__, type); in aarch64_insn_gen_data2()
1041 pr_err("%s: unknown variant encoding %d\n", __func__, variant); in aarch64_insn_gen_data2()
1069 pr_err("%s: unknown data3 encoding %d\n", __func__, type); in aarch64_insn_gen_data3()
1080 pr_err("%s: unknown variant encoding %d\n", __func__, variant); in aarch64_insn_gen_data3()
1130 pr_err("%s: unknown logical encoding %d\n", __func__, type); in aarch64_insn_gen_logical_shifted_reg()
1137 pr_err("%s: invalid shift encoding %d\n", __func__, in aarch64_insn_gen_logical_shifted_reg()
1145 pr_err("%s: invalid shift encoding %d\n", __func__, in aarch64_insn_gen_logical_shifted_reg()
1151 pr_err("%s: unknown variant encoding %d\n", __func__, variant); in aarch64_insn_gen_logical_shifted_reg()
1195 pr_err("%s: unknown adr encoding %d\n", __func__, type); in aarch64_insn_gen_adr()
1331 pr_err("%s: unknown variant encoding %d\n", __func__, variant); in aarch64_encode_immediate()
1433 pr_err("%s: unknown logical encoding %d\n", __func__, type); in aarch64_insn_gen_logical_immediate()
1464 pr_err("%s: unknown variant encoding %d\n", __func__, variant); in aarch64_insn_gen_extr()
1508 pr_err("%s: unknown dmb type %d\n", __func__, type); in aarch64_insn_gen_dmb()