1 /* 2 * xtensa/coreasm.h -- assembler-specific definitions that depend on CORE configuration 3 * 4 * Source for configuration-independent binaries (which link in a 5 * configuration-specific HAL library) must NEVER include this file. 6 * It is perfectly normal, however, for the HAL itself to include this file. 7 * 8 * This file must NOT include xtensa/config/system.h. Any assembler 9 * header file that depends on system information should likely go 10 * in a new systemasm.h (or sysasm.h) header file. 11 * 12 * NOTE: macro beqi32 is NOT configuration-dependent, and is placed 13 * here until we have a proper configuration-independent header file. 14 */ 15 16 /* $Id: //depot/rel/Foxhill/dot.9/Xtensa/OS/include/xtensa/coreasm.h#1 $ */ 17 18 /* 19 * Copyright (c) 2000-2014 Tensilica Inc. 20 * 21 * Permission is hereby granted, free of charge, to any person obtaining 22 * a copy of this software and associated documentation files (the 23 * "Software"), to deal in the Software without restriction, including 24 * without limitation the rights to use, copy, modify, merge, publish, 25 * distribute, sublicense, and/or sell copies of the Software, and to 26 * permit persons to whom the Software is furnished to do so, subject to 27 * the following conditions: 28 * 29 * The above copyright notice and this permission notice shall be included 30 * in all copies or substantial portions of the Software. 31 * 32 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 33 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 34 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. 35 * IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY 36 * CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, 37 * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE 38 * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. 39 */ 40 41 #ifndef XTENSA_COREASM_H 42 #define XTENSA_COREASM_H 43 44 /* 45 * Tell header files this is assembly source, so they can avoid non-assembler 46 * definitions (eg. C types etc): 47 */ 48 #ifndef _ASMLANGUAGE /* conditionalize to avoid cpp warnings (3rd parties might use same macro) */ 49 #define _ASMLANGUAGE 50 #endif 51 52 #include <xtensa/config/core.h> 53 #include <xtensa/config/specreg.h> 54 #include <xtensa/config/system.h> 55 56 /* 57 * Assembly-language specific definitions (assembly macros, etc.). 58 */ 59 60 /*---------------------------------------------------------------------- 61 * find_ms_setbit 62 * 63 * This macro finds the most significant bit that is set in <as> 64 * and return its index + <base> in <ad>, or <base> - 1 if <as> is zero. 65 * The index counts starting at zero for the lsbit, so the return 66 * value ranges from <base>-1 (no bit set) to <base>+31 (msbit set). 67 * 68 * Parameters: 69 * <ad> destination address register (any register) 70 * <as> source address register 71 * <at> temporary address register (must be different than <as>) 72 * <base> constant value added to result (usually 0 or 1) 73 * On entry: 74 * <ad> = undefined if different than <as> 75 * <as> = value whose most significant set bit is to be found 76 * <at> = undefined 77 * no other registers are used by this macro. 78 * On exit: 79 * <ad> = <base> + index of msbit set in original <as>, 80 * = <base> - 1 if original <as> was zero. 81 * <as> clobbered (if not <ad>) 82 * <at> clobbered (if not <ad>) 83 * Example: 84 * find_ms_setbit a0, a4, a0, 0 -- return in a0 index of msbit set in a4 85 */ 86 87 .macro find_ms_setbit ad, as, at, base 88 #if XCHAL_HAVE_NSA 89 movi \at, 31+\base 90 nsau \as, \as // get index of \as, numbered from msbit (32 if absent) 91 sub \ad, \at, \as // get numbering from lsbit (0..31, -1 if absent) 92 #else /* XCHAL_HAVE_NSA */ 93 movi \at, \base // start with result of 0 (point to lsbit of 32) 94 95 beqz \as, 2f // special case for zero argument: return -1 96 bltui \as, 0x10000, 1f // is it one of the 16 lsbits? (if so, check lower 16 bits) 97 addi \at, \at, 16 // no, increment result to upper 16 bits (of 32) 98 //srli \as, \as, 16 // check upper half (shift right 16 bits) 99 extui \as, \as, 16, 16 // check upper half (shift right 16 bits) 100 1: bltui \as, 0x100, 1f // is it one of the 8 lsbits? (if so, check lower 8 bits) 101 addi \at, \at, 8 // no, increment result to upper 8 bits (of 16) 102 srli \as, \as, 8 // shift right to check upper 8 bits 103 1: bltui \as, 0x10, 1f // is it one of the 4 lsbits? (if so, check lower 4 bits) 104 addi \at, \at, 4 // no, increment result to upper 4 bits (of 8) 105 srli \as, \as, 4 // shift right 4 bits to check upper half 106 1: bltui \as, 0x4, 1f // is it one of the 2 lsbits? (if so, check lower 2 bits) 107 addi \at, \at, 2 // no, increment result to upper 2 bits (of 4) 108 srli \as, \as, 2 // shift right 2 bits to check upper half 109 1: bltui \as, 0x2, 1f // is it the lsbit? 110 addi \at, \at, 2 // no, increment result to upper bit (of 2) 111 2: addi \at, \at, -1 // (from just above: add 1; from beqz: return -1) 112 //srli \as, \as, 1 113 1: // done! \at contains index of msbit set (or -1 if none set) 114 .if 0x\ad - 0x\at // destination different than \at ? (works because regs are a0-a15) 115 mov \ad, \at // then move result to \ad 116 .endif 117 #endif /* XCHAL_HAVE_NSA */ 118 .endm // find_ms_setbit 119 120 /*---------------------------------------------------------------------- 121 * find_ls_setbit 122 * 123 * This macro finds the least significant bit that is set in <as>, 124 * and return its index in <ad>. 125 * Usage is the same as for the find_ms_setbit macro. 126 * Example: 127 * find_ls_setbit a0, a4, a0, 0 -- return in a0 index of lsbit set in a4 128 */ 129 130 .macro find_ls_setbit ad, as, at, base 131 neg \at, \as // keep only the least-significant bit that is set... 132 and \as, \at, \as // ... in \as 133 find_ms_setbit \ad, \as, \at, \base 134 .endm // find_ls_setbit 135 136 /*---------------------------------------------------------------------- 137 * find_ls_one 138 * 139 * Same as find_ls_setbit with base zero. 140 * Source (as) and destination (ad) registers must be different. 141 * Provided for backward compatibility. 142 */ 143 144 .macro find_ls_one ad, as 145 find_ls_setbit \ad, \as, \ad, 0 146 .endm // find_ls_one 147 148 /*---------------------------------------------------------------------- 149 * floop, floopnez, floopgtz, floopend 150 * 151 * These macros are used for fast inner loops that 152 * work whether or not the Loops options is configured. 153 * If the Loops option is configured, they simply use 154 * the zero-overhead LOOP instructions; otherwise 155 * they use explicit decrement and branch instructions. 156 * 157 * They are used in pairs, with floop, floopnez or floopgtz 158 * at the beginning of the loop, and floopend at the end. 159 * 160 * Each pair of loop macro calls must be given the loop count 161 * address register and a unique label for that loop. 162 * 163 * Example: 164 * 165 * movi a3, 16 // loop 16 times 166 * floop a3, myloop1 167 * : 168 * bnez a7, end1 // exit loop if a7 != 0 169 * : 170 * floopend a3, myloop1 171 * end1: 172 * 173 * Like the LOOP instructions, these macros cannot be 174 * nested, must include at least one instruction, 175 * cannot call functions inside the loop, etc. 176 * The loop can be exited by jumping to the instruction 177 * following floopend (or elsewhere outside the loop), 178 * or continued by jumping to a NOP instruction placed 179 * immediately before floopend. 180 * 181 * Unlike LOOP instructions, the register passed to floop* 182 * cannot be used inside the loop, because it is used as 183 * the loop counter if the Loops option is not configured. 184 * And its value is undefined after exiting the loop. 185 * And because the loop counter register is active inside 186 * the loop, you can't easily use this construct to loop 187 * across a register file using ROTW as you might with LOOP 188 * instructions, unless you copy the loop register along. 189 */ 190 191 /* Named label version of the macros: */ 192 193 .macro floop ar, endlabel 194 floop_ \ar, .Lfloopstart_\endlabel, .Lfloopend_\endlabel 195 .endm 196 197 .macro floopnez ar, endlabel 198 floopnez_ \ar, .Lfloopstart_\endlabel, .Lfloopend_\endlabel 199 .endm 200 201 .macro floopgtz ar, endlabel 202 floopgtz_ \ar, .Lfloopstart_\endlabel, .Lfloopend_\endlabel 203 .endm 204 205 .macro floopend ar, endlabel 206 floopend_ \ar, .Lfloopstart_\endlabel, .Lfloopend_\endlabel 207 .endm 208 209 /* Numbered local label version of the macros: */ 210 #if 0 /*UNTESTED*/ 211 .macro floop89 ar 212 floop_ \ar, 8, 9f 213 .endm 214 215 .macro floopnez89 ar 216 floopnez_ \ar, 8, 9f 217 .endm 218 219 .macro floopgtz89 ar 220 floopgtz_ \ar, 8, 9f 221 .endm 222 223 .macro floopend89 ar 224 floopend_ \ar, 8b, 9 225 .endm 226 #endif /*0*/ 227 228 /* Underlying version of the macros: */ 229 230 .macro floop_ ar, startlabel, endlabelref 231 .ifdef _infloop_ 232 .if _infloop_ 233 .err // Error: floop cannot be nested 234 .endif 235 .endif 236 .set _infloop_, 1 237 #if XCHAL_HAVE_LOOPS 238 loop \ar, \endlabelref 239 #else /* XCHAL_HAVE_LOOPS */ 240 \startlabel: 241 addi \ar, \ar, -1 242 #endif /* XCHAL_HAVE_LOOPS */ 243 .endm // floop_ 244 245 .macro floopnez_ ar, startlabel, endlabelref 246 .ifdef _infloop_ 247 .if _infloop_ 248 .err // Error: floopnez cannot be nested 249 .endif 250 .endif 251 .set _infloop_, 1 252 #if XCHAL_HAVE_LOOPS 253 loopnez \ar, \endlabelref 254 #else /* XCHAL_HAVE_LOOPS */ 255 beqz \ar, \endlabelref 256 \startlabel: 257 addi \ar, \ar, -1 258 #endif /* XCHAL_HAVE_LOOPS */ 259 .endm // floopnez_ 260 261 .macro floopgtz_ ar, startlabel, endlabelref 262 .ifdef _infloop_ 263 .if _infloop_ 264 .err // Error: floopgtz cannot be nested 265 .endif 266 .endif 267 .set _infloop_, 1 268 #if XCHAL_HAVE_LOOPS 269 loopgtz \ar, \endlabelref 270 #else /* XCHAL_HAVE_LOOPS */ 271 bltz \ar, \endlabelref 272 beqz \ar, \endlabelref 273 \startlabel: 274 addi \ar, \ar, -1 275 #endif /* XCHAL_HAVE_LOOPS */ 276 .endm // floopgtz_ 277 278 279 .macro floopend_ ar, startlabelref, endlabel 280 .ifndef _infloop_ 281 .err // Error: floopend without matching floopXXX 282 .endif 283 .ifeq _infloop_ 284 .err // Error: floopend without matching floopXXX 285 .endif 286 .set _infloop_, 0 287 #if ! XCHAL_HAVE_LOOPS 288 bnez \ar, \startlabelref 289 #endif /* XCHAL_HAVE_LOOPS */ 290 \endlabel: 291 .endm // floopend_ 292 293 /*---------------------------------------------------------------------- 294 * crsil -- conditional RSIL (read/set interrupt level) 295 * 296 * Executes the RSIL instruction if it exists, else just reads PS. 297 * The RSIL instruction does not exist in the new exception architecture 298 * if the interrupt option is not selected. 299 */ 300 301 .macro crsil ar, newlevel 302 #if XCHAL_HAVE_OLD_EXC_ARCH || XCHAL_HAVE_INTERRUPTS 303 rsil \ar, \newlevel 304 #else 305 rsr.ps \ar 306 #endif 307 .endm // crsil 308 309 /*---------------------------------------------------------------------- 310 * safe_movi_a0 -- move constant into a0 when L32R is not safe 311 * 312 * This macro is typically used by interrupt/exception handlers. 313 * Loads a 32-bit constant in a0, without using any other register, 314 * and without corrupting the LITBASE register, even when the 315 * value of the LITBASE register is unknown (eg. when application 316 * code and interrupt/exception handling code are built independently, 317 * and thus with independent values of the LITBASE register; 318 * debug monitors are one example of this). 319 * 320 * Worst-case size of resulting code: 17 bytes. 321 */ 322 323 .macro safe_movi_a0 constant 324 #if XCHAL_HAVE_ABSOLUTE_LITERALS 325 /* Contort a PC-relative literal load even though we may be in litbase-relative mode: */ 326 j 1f 327 .begin no-transform // ensure what follows is assembled exactly as-is 328 .align 4 // ensure constant and call0 target ... 329 .byte 0 // ... are 4-byte aligned (call0 instruction is 3 bytes long) 330 1: call0 2f // read PC (that follows call0) in a0 331 .long \constant // 32-bit constant to load into a0 332 2: 333 .end no-transform 334 l32i a0, a0, 0 // load constant 335 #else 336 movi a0, \constant // no LITBASE, can assume PC-relative L32R 337 #endif 338 .endm 339 340 341 342 343 /*---------------------------------------------------------------------- 344 * window_spill{4,8,12} 345 * 346 * These macros spill callers' register windows to the stack. 347 * They work for both privileged and non-privileged tasks. 348 * Must be called from a windowed ABI context, eg. within 349 * a windowed ABI function (ie. valid stack frame, window 350 * exceptions enabled, not in exception mode, etc). 351 * 352 * This macro requires a single invocation of the window_spill_common 353 * macro in the same assembly unit and section. 354 * 355 * Note that using window_spill{4,8,12} macros is more efficient 356 * than calling a function implemented using window_spill_function, 357 * because the latter needs extra code to figure out the size of 358 * the call to the spilling function. 359 * 360 * Example usage: 361 * 362 * .text 363 * .align 4 364 * .global some_function 365 * .type some_function,@function 366 * some_function: 367 * entry a1, 16 368 * : 369 * : 370 * 371 * window_spill4 // Spill windows of some_function's callers; preserves a0..a3 only; 372 * // to use window_spill{8,12} in this example function we'd have 373 * // to increase space allocated by the entry instruction, because 374 * // 16 bytes only allows call4; 32 or 48 bytes (+locals) are needed 375 * // for call8/window_spill8 or call12/window_spill12 respectively. 376 * 377 * : 378 * 379 * retw 380 * 381 * window_spill_common // instantiates code used by window_spill4 382 * 383 * 384 * On entry: 385 * none (if window_spill4) 386 * stack frame has enough space allocated for call8 (if window_spill8) 387 * stack frame has enough space allocated for call12 (if window_spill12) 388 * On exit: 389 * a4..a15 clobbered (if window_spill4) 390 * a8..a15 clobbered (if window_spill8) 391 * a12..a15 clobbered (if window_spill12) 392 * no caller windows are in live registers 393 */ 394 395 .macro window_spill4 396 #if XCHAL_HAVE_WINDOWED 397 # if XCHAL_NUM_AREGS == 16 398 movi a15, 0 // for 16-register files, no need to call to reach the end 399 # elif XCHAL_NUM_AREGS == 32 400 call4 .L__wdwspill_assist28 // call deep enough to clear out any live callers 401 # elif XCHAL_NUM_AREGS == 64 402 call4 .L__wdwspill_assist60 // call deep enough to clear out any live callers 403 # endif 404 #endif 405 .endm // window_spill4 406 407 .macro window_spill8 408 #if XCHAL_HAVE_WINDOWED 409 # if XCHAL_NUM_AREGS == 16 410 movi a15, 0 // for 16-register files, no need to call to reach the end 411 # elif XCHAL_NUM_AREGS == 32 412 call8 .L__wdwspill_assist24 // call deep enough to clear out any live callers 413 # elif XCHAL_NUM_AREGS == 64 414 call8 .L__wdwspill_assist56 // call deep enough to clear out any live callers 415 # endif 416 #endif 417 .endm // window_spill8 418 419 .macro window_spill12 420 #if XCHAL_HAVE_WINDOWED 421 # if XCHAL_NUM_AREGS == 16 422 movi a15, 0 // for 16-register files, no need to call to reach the end 423 # elif XCHAL_NUM_AREGS == 32 424 call12 .L__wdwspill_assist20 // call deep enough to clear out any live callers 425 # elif XCHAL_NUM_AREGS == 64 426 call12 .L__wdwspill_assist52 // call deep enough to clear out any live callers 427 # endif 428 #endif 429 .endm // window_spill12 430 431 432 /*---------------------------------------------------------------------- 433 * window_spill_function 434 * 435 * This macro outputs a function that will spill its caller's callers' 436 * register windows to the stack. Eg. it could be used to implement 437 * a version of xthal_window_spill() that works in non-privileged tasks. 438 * This works for both privileged and non-privileged tasks. 439 * 440 * Typical usage: 441 * 442 * .text 443 * .align 4 444 * .global my_spill_function 445 * .type my_spill_function,@function 446 * my_spill_function: 447 * window_spill_function 448 * 449 * On entry to resulting function: 450 * none 451 * On exit from resulting function: 452 * none (no caller windows are in live registers) 453 */ 454 455 .macro window_spill_function 456 #if XCHAL_HAVE_WINDOWED 457 # if XCHAL_NUM_AREGS == 32 458 entry sp, 48 459 bbci.l a0, 31, 1f // branch if called with call4 460 bbsi.l a0, 30, 2f // branch if called with call12 461 call8 .L__wdwspill_assist16 // called with call8, only need another 8 462 retw 463 1: call12 .L__wdwspill_assist16 // called with call4, only need another 12 464 retw 465 2: call4 .L__wdwspill_assist16 // called with call12, only need another 4 466 retw 467 # elif XCHAL_NUM_AREGS == 64 468 entry sp, 48 469 bbci.l a0, 31, 1f // branch if called with call4 470 bbsi.l a0, 30, 2f // branch if called with call12 471 call4 .L__wdwspill_assist52 // called with call8, only need a call4 472 retw 473 1: call8 .L__wdwspill_assist52 // called with call4, only need a call8 474 retw 475 2: call12 .L__wdwspill_assist40 // called with call12, can skip a call12 476 retw 477 # elif XCHAL_NUM_AREGS == 16 478 entry sp, 16 479 bbci.l a0, 31, 1f // branch if called with call4 480 bbsi.l a0, 30, 2f // branch if called with call12 481 movi a7, 0 // called with call8 482 retw 483 1: movi a11, 0 // called with call4 484 2: retw // if called with call12, everything already spilled 485 486 // movi a15, 0 // trick to spill all but the direct caller 487 // j 1f 488 // // The entry instruction is magical in the assembler (gets auto-aligned) 489 // // so we have to jump to it to avoid falling through the padding. 490 // // We need entry/retw to know where to return. 491 //1: entry sp, 16 492 // retw 493 # else 494 # error "unrecognized address register file size" 495 # endif 496 497 #endif /* XCHAL_HAVE_WINDOWED */ 498 window_spill_common 499 .endm // window_spill_function 500 501 /*---------------------------------------------------------------------- 502 * window_spill_common 503 * 504 * Common code used by any number of invocations of the window_spill## 505 * and window_spill_function macros. 506 * 507 * Must be instantiated exactly once within a given assembly unit, 508 * within call/j range of and same section as window_spill## 509 * macro invocations for that assembly unit. 510 * (Is automatically instantiated by the window_spill_function macro.) 511 */ 512 513 .macro window_spill_common 514 #if XCHAL_HAVE_WINDOWED && (XCHAL_NUM_AREGS == 32 || XCHAL_NUM_AREGS == 64) 515 .ifndef .L__wdwspill_defined 516 # if XCHAL_NUM_AREGS >= 64 517 .L__wdwspill_assist60: 518 entry sp, 32 519 call8 .L__wdwspill_assist52 520 retw 521 .L__wdwspill_assist56: 522 entry sp, 16 523 call4 .L__wdwspill_assist52 524 retw 525 .L__wdwspill_assist52: 526 entry sp, 48 527 call12 .L__wdwspill_assist40 528 retw 529 .L__wdwspill_assist40: 530 entry sp, 48 531 call12 .L__wdwspill_assist28 532 retw 533 # endif 534 .L__wdwspill_assist28: 535 entry sp, 48 536 call12 .L__wdwspill_assist16 537 retw 538 .L__wdwspill_assist24: 539 entry sp, 32 540 call8 .L__wdwspill_assist16 541 retw 542 .L__wdwspill_assist20: 543 entry sp, 16 544 call4 .L__wdwspill_assist16 545 retw 546 .L__wdwspill_assist16: 547 entry sp, 16 548 movi a15, 0 549 retw 550 .set .L__wdwspill_defined, 1 551 .endif 552 #endif /* XCHAL_HAVE_WINDOWED with 32 or 64 aregs */ 553 .endm // window_spill_common 554 555 /*---------------------------------------------------------------------- 556 * beqi32 557 * 558 * macro implements version of beqi for arbitrary 32-bit immediate value 559 * 560 * beqi32 ax, ay, imm32, label 561 * 562 * Compares value in register ax with imm32 value and jumps to label if 563 * equal. Clobbers register ay if needed 564 * 565 */ 566 .macro beqi32 ax, ay, imm, label 567 .ifeq ((\imm-1) & ~7) // 1..8 ? 568 beqi \ax, \imm, \label 569 .else 570 .ifeq (\imm+1) // -1 ? 571 beqi \ax, \imm, \label 572 .else 573 .ifeq (\imm) // 0 ? 574 beqz \ax, \label 575 .else 576 // We could also handle immediates 10,12,16,32,64,128,256 577 // but it would be a long macro... 578 movi \ay, \imm 579 beq \ax, \ay, \label 580 .endif 581 .endif 582 .endif 583 .endm // beqi32 584 585 /*---------------------------------------------------------------------- 586 * isync_retw_nop 587 * 588 * This macro must be invoked immediately after ISYNC if ISYNC 589 * would otherwise be immediately followed by RETW (or other instruction 590 * modifying WindowBase or WindowStart), in a context where 591 * kernel vector mode may be selected, and level-one interrupts 592 * and window overflows may be enabled, on an XEA1 configuration. 593 * 594 * On hardware with erratum "XEA1KWIN" (see <xtensa/core.h> for details), 595 * XEA1 code must have at least one instruction between ISYNC and RETW if 596 * run in kernel vector mode with interrupts and window overflows enabled. 597 */ 598 .macro isync_retw_nop 599 #if XCHAL_MAYHAVE_ERRATUM_XEA1KWIN 600 nop 601 #endif 602 .endm 603 604 /*---------------------------------------------------------------------- 605 * isync_return_nop 606 * 607 * This macro should be used instead of isync_retw_nop in code that is 608 * intended to run on both the windowed and call0 ABIs 609 */ 610 .macro isync_return_nop 611 #ifdef __XTENSA_WINDOWED_ABI__ 612 isync_retw_nop 613 #endif 614 .endm 615 616 /*---------------------------------------------------------------------- 617 * isync_erratum453 618 * 619 * This macro must be invoked at certain points in the code, 620 * such as in exception and interrupt vectors in particular, 621 * to work around erratum 453. 622 */ 623 .macro isync_erratum453 624 #if XCHAL_ERRATUM_453 625 isync 626 #endif 627 .endm 628 629 630 /*---------------------------------------------------------------------- 631 * readsr 632 * 633 * wrapper for 'rsr' that constructs register names that involve levels 634 * e.g. EPCn etc. Use like so: 635 * readsr epc XCHAL_DEBUGLEVEL a2 636 */ 637 .macro readsr reg suf ar 638 rsr.\reg\suf \ar 639 .endm 640 641 /*---------------------------------------------------------------------- 642 * writesr 643 * 644 * wrapper for 'wsr' that constructs register names that involve levels 645 * e.g. EPCn etc. Use like so: 646 * writesr epc XCHAL_DEBUGLEVEL a2 647 */ 648 .macro writesr reg suf ar 649 wsr.\reg\suf \ar 650 .endm 651 652 /*---------------------------------------------------------------------- 653 * xchgsr 654 * 655 * wrapper for 'xsr' that constructs register names that involve levels 656 * e.g. EPCn etc. Use like so: 657 * xchgsr epc XCHAL_DEBUGLEVEL a2 658 */ 659 .macro xchgsr reg suf ar 660 xsr.\reg\suf \ar 661 .endm 662 663 /*---------------------------------------------------------------------- 664 * INDEX_SR 665 * 666 * indexing wrapper for rsr/wsr/xsr that constructs register names from 667 * the provided base name and the current index. Use like so: 668 * .set _idx, 0 669 * INDEX_SR rsr.ccompare a2 670 * 671 * this yields: rsr.ccompare0 a2 672 */ 673 .macro INDEX_SR instr ar 674 .ifeq (_idx) 675 &instr&0 \ar 676 .endif 677 .ifeq (_idx-1) 678 &instr&1 \ar 679 .endif 680 .ifeq (_idx-2) 681 &instr&2 \ar 682 .endif 683 .ifeq (_idx-3) 684 &instr&3 \ar 685 .endif 686 .ifeq (_idx-4) 687 &instr&4 \ar 688 .endif 689 .ifeq (_idx-5) 690 &instr&5 \ar 691 .endif 692 .ifeq (_idx-6) 693 &instr&6 \ar 694 .endif 695 .ifeq (_idx-7) 696 &instr&7 \ar 697 .endif 698 .endm 699 700 701 /*---------------------------------------------------------------------- 702 * abs 703 * 704 * implements abs on machines that do not have it configured 705 */ 706 707 #if !XCHAL_HAVE_ABS 708 .macro abs arr, ars 709 .ifc \arr, \ars 710 //src equal dest is less efficient 711 bgez \arr, 1f 712 neg \arr, \arr 713 1: 714 .else 715 neg \arr, \ars 716 movgez \arr, \ars, \ars 717 .endif 718 .endm 719 #endif /* !XCHAL_HAVE_ABS */ 720 721 722 /*---------------------------------------------------------------------- 723 * addx2 724 * 725 * implements addx2 on machines that do not have it configured 726 * 727 */ 728 729 #if !XCHAL_HAVE_ADDX 730 .macro addx2 arr, ars, art 731 .ifc \arr, \art 732 .ifc \arr, \ars 733 // addx2 a, a, a (not common) 734 .err 735 .else 736 add \arr, \ars, \art 737 add \arr, \ars, \art 738 .endif 739 .else 740 //addx2 a, b, c 741 //addx2 a, a, b 742 //addx2 a, b, b 743 slli \arr, \ars, 1 744 add \arr, \arr, \art 745 .endif 746 .endm 747 #endif /* !XCHAL_HAVE_ADDX */ 748 749 /*---------------------------------------------------------------------- 750 * addx4 751 * 752 * implements addx4 on machines that do not have it configured 753 * 754 */ 755 756 #if !XCHAL_HAVE_ADDX 757 .macro addx4 arr, ars, art 758 .ifc \arr, \art 759 .ifc \arr, \ars 760 // addx4 a, a, a (not common) 761 .err 762 .else 763 //# addx4 a, b, a 764 add \arr, \ars, \art 765 add \arr, \ars, \art 766 add \arr, \ars, \art 767 add \arr, \ars, \art 768 .endif 769 .else 770 //addx4 a, b, c 771 //addx4 a, a, b 772 //addx4 a, b, b 773 slli \arr, \ars, 2 774 add \arr, \arr, \art 775 .endif 776 .endm 777 #endif /* !XCHAL_HAVE_ADDX */ 778 779 /*---------------------------------------------------------------------- 780 * addx8 781 * 782 * implements addx8 on machines that do not have it configured 783 * 784 */ 785 786 #if !XCHAL_HAVE_ADDX 787 .macro addx8 arr, ars, art 788 .ifc \arr, \art 789 .ifc \arr, \ars 790 //addx8 a, a, a (not common) 791 .err 792 .else 793 //addx8 a, b, a 794 add \arr, \ars, \art 795 add \arr, \ars, \art 796 add \arr, \ars, \art 797 add \arr, \ars, \art 798 add \arr, \ars, \art 799 add \arr, \ars, \art 800 add \arr, \ars, \art 801 add \arr, \ars, \art 802 .endif 803 .else 804 //addx8 a, b, c 805 //addx8 a, a, b 806 //addx8 a, b, b 807 slli \arr, \ars, 3 808 add \arr, \arr, \art 809 .endif 810 .endm 811 #endif /* !XCHAL_HAVE_ADDX */ 812 813 814 /*---------------------------------------------------------------------- 815 * rfe_rfue 816 * 817 * Maps to RFUE on XEA1, and RFE on XEA2. No mapping on XEAX. 818 */ 819 820 #if XCHAL_HAVE_XEA1 821 .macro rfe_rfue 822 rfue 823 .endm 824 #elif XCHAL_HAVE_XEA2 825 .macro rfe_rfue 826 rfe 827 .endm 828 #endif 829 830 831 /*---------------------------------------------------------------------- 832 * abi_entry 833 * 834 * Generate proper function entry sequence for the current ABI 835 * (windowed or call0). Takes care of allocating stack space (up to 1kB) 836 * and saving the return PC, if necessary. The corresponding abi_return 837 * macro does the corresponding stack deallocation and restoring return PC. 838 * 839 * Parameters are: 840 * 841 * locsize Number of bytes to allocate on the stack 842 * for local variables (and for args to pass to 843 * callees, if any calls are made). Defaults to zero. 844 * The macro rounds this up to a multiple of 16. 845 * NOTE: large values are allowed (e.g. up to 1 GB). 846 * 847 * callsize Maximum call size made by this function. 848 * Leave zero (default) for leaf functions, i.e. if 849 * this function makes no calls to other functions. 850 * Otherwise must be set to 4, 8, or 12 according 851 * to whether the "largest" call made is a call[x]4, 852 * call[x]8, or call[x]12 (for call0 ABI, it makes 853 * no difference whether this is set to 4, 8 or 12, 854 * but it must be set to one of these values). 855 * 856 * NOTE: It is up to the caller to align the entry point, declare the 857 * function symbol, make it global, etc. 858 * 859 * NOTE: This macro relies on assembler relaxation for large values 860 * of locsize. It might not work with the no-transform directive. 861 * NOTE: For the call0 ABI, this macro ensures SP is allocated or 862 * de-allocated cleanly, i.e. without temporarily allocating too much 863 * (or allocating negatively!) due to addi relaxation. 864 * 865 * NOTE: Generating the proper sequence and register allocation for 866 * making calls in an ABI independent manner is a separate topic not 867 * covered by this macro. 868 * 869 * NOTE: To access arguments, you can't use a fixed offset from SP. 870 * The offset depends on the ABI, whether the function is leaf, etc. 871 * The simplest method is probably to use the .locsz symbol, which 872 * is set by this macro to the actual number of bytes allocated on 873 * the stack, in other words, to the offset from SP to the arguments. 874 * E.g. for a function whose arguments are all 32-bit integers, you 875 * can get the 7th and 8th arguments (1st and 2nd args stored on stack) 876 * using: 877 * l32i a2, sp, .locsz 878 * l32i a3, sp, .locsz+4 879 * (this example works as long as locsize is under L32I's offset limit 880 * of 1020 minus up to 48 bytes of ABI-specific stack usage; 881 * otherwise you might first need to do "addi a?, sp, .locsz" 882 * or similar sequence). 883 * 884 * NOTE: For call0 ABI, this macro (and abi_return) may clobber a9 885 * (a caller-saved register). 886 * 887 * Examples: 888 * abi_entry 889 * abi_entry 5 890 * abi_entry 22, 8 891 * abi_entry 0, 4 892 */ 893 894 /* 895 * Compute .locsz and .callsz without emitting any instructions. 896 * Used by both abi_entry and abi_return. 897 * Assumes locsize >= 0. 898 */ 899 .macro abi_entry_size locsize=0, callsize=0 900 #if XCHAL_HAVE_WINDOWED && !__XTENSA_CALL0_ABI__ 901 .ifeq \callsize 902 .set .callsz, 16 903 .else 904 .ifeq \callsize-4 905 .set .callsz, 16 906 .else 907 .ifeq \callsize-8 908 .set .callsz, 32 909 .else 910 .ifeq \callsize-12 911 .set .callsz, 48 912 .else 913 .error "abi_entry: invalid call size \callsize" 914 .endif 915 .endif 916 .endif 917 .endif 918 .set .locsz, .callsz + ((\locsize + 15) & -16) 919 #else 920 .set .callsz, \callsize 921 .if .callsz /* if calls, need space for return PC */ 922 .set .locsz, (\locsize + 4 + 15) & -16 923 .else 924 .set .locsz, (\locsize + 15) & -16 925 .endif 926 #endif 927 .endm 928 929 .macro abi_entry locsize=0, callsize=0 930 .iflt \locsize 931 .error "abi_entry: invalid negative size of locals (\locsize)" 932 .endif 933 abi_entry_size \locsize, \callsize 934 #if XCHAL_HAVE_WINDOWED && !__XTENSA_CALL0_ABI__ 935 # define ABI_ENTRY_MINSIZE 3 /* size of abi_entry (no arguments) instructions in bytes */ 936 .ifgt .locsz - 32760 /* .locsz > 32760 (ENTRY's max range)? */ 937 /* Funky computation to try to have assembler use addmi efficiently if possible: */ 938 entry sp, 0x7F00 + (.locsz & 0xF0) 939 addi a12, sp, - ((.locsz & -0x100) - 0x7F00) 940 movsp sp, a12 941 .else 942 entry sp, .locsz 943 .endif 944 #else 945 # define ABI_ENTRY_MINSIZE 0 /* size of abi_entry (no arguments) instructions in bytes */ 946 .if .locsz 947 .ifle .locsz - 128 /* if locsz <= 128 */ 948 addi sp, sp, -.locsz 949 .if .callsz 950 s32i a0, sp, .locsz - 4 951 .endif 952 .elseif .callsz /* locsz > 128, with calls: */ 953 movi a9, .locsz - 16 /* note: a9 is caller-saved */ 954 addi sp, sp, -16 955 s32i a0, sp, 12 956 sub sp, sp, a9 957 .else /* locsz > 128, no calls: */ 958 movi a9, .locsz 959 sub sp, sp, a9 960 .endif /* end */ 961 .endif 962 #endif 963 .endm 964 965 966 967 /*---------------------------------------------------------------------- 968 * abi_return 969 * 970 * Generate proper function exit sequence for the current ABI 971 * (windowed or call0). Takes care of freeing stack space and 972 * restoring the return PC, if necessary. 973 * NOTE: This macro MUST be invoked following a corresponding 974 * abi_entry macro invocation. For call0 ABI in particular, 975 * all stack and PC restoration are done according to the last 976 * abi_entry macro invoked before this macro in the assembly file. 977 * 978 * Normally this macro takes no arguments. However to allow 979 * for placing abi_return *before* abi_entry (as must be done 980 * for some highly optimized assembly), it optionally takes 981 * exactly the same arguments as abi_entry. 982 */ 983 984 .macro abi_return locsize=-1, callsize=0 985 .ifge \locsize 986 abi_entry_size \locsize, \callsize 987 .endif 988 #if XCHAL_HAVE_WINDOWED && !__XTENSA_CALL0_ABI__ 989 retw 990 #else 991 .if .locsz 992 .iflt .locsz - 128 /* if locsz < 128 */ 993 .if .callsz 994 l32i a0, sp, .locsz - 4 995 .endif 996 addi sp, sp, .locsz 997 .elseif .callsz /* locsz >= 128, with calls: */ 998 addi a9, sp, .locsz - 16 999 l32i a0, a9, 12 1000 addi sp, a9, 16 1001 .else /* locsz >= 128, no calls: */ 1002 movi a9, .locsz 1003 add sp, sp, a9 1004 .endif /* end */ 1005 .endif 1006 ret 1007 #endif 1008 .endm 1009 1010 1011 /* 1012 * HW erratum fixes. 1013 */ 1014 1015 .macro hw_erratum_487_fix 1016 #if defined XSHAL_ERRATUM_487_FIX 1017 isync 1018 #endif 1019 .endm 1020 1021 /* 1022 * These macros are internal, subject to change, and should not be used in 1023 * any new code. 1024 */ 1025 1026 #define _GBL(x) .global x 1027 #define _TYP(x) .type x,@function 1028 #define _ALN(x) .align x 1029 #define _SIZ(x) .size x, . - x 1030 #define _MKEND(x) .purgem endfunc ; .macro endfunc ; _SIZ(x) ; .purgem endfunc ; .macro endfunc ; .endm ; .endm 1031 #define _SYMT(x) _GBL(x); _MKEND(x); _TYP(x); _ALN(4); x: 1032 #define _SYM2(x) _GBL(x); _TYP(x); x: 1033 #define _SYM(x) _GBL(x); _MKEND(x); _ALN(4); x: 1034 .macro endfunc ; .endm 1035 1036 /* 1037 * the DECLFUNC() macro provides a mechanism for implementing both the 1038 * standard and _nw interface with a single copy of the code. 1039 * 1040 * For Call0 ABI there is one function definition which is labeled with 1041 * both the xthal_..._nw and xthal_... symbols. 1042 * 1043 * For windowed ABI, two compilations are involved (one with the __NW_FUNCTION__ 1044 * symbol defined) resulting in two separate functions (the _nw one without 1045 * the window adjustments). 1046 */ 1047 1048 #if defined(__NW_FUNCTION__) 1049 # define DECLFUNC(x) _SYMT(x ## _nw) 1050 #else 1051 # if defined (__XTENSA_CALL0_ABI__) 1052 # define DECLFUNC(x) _SYMT(x); _SYM2(x ## _nw) 1053 # else 1054 # define DECLFUNC(x) _SYMT(x) 1055 # endif 1056 #endif 1057 1058 #endif /*XTENSA_COREASM_H*/ 1059