Lines Matching refs:mov

65 	mov loc0=rp
67 mov out0=in0 // filename
69 mov out1=in1 // argv
70 mov out2=in2 // envp
74 mov ar.pfs=loc1 // restore ar.pfs
78 mov rp=loc0
79 (p6) mov ar.pfs=r0 // clear ar.pfs on success
88 mov ar.unat=0; mov ar.lc=0
89 mov r4=0; mov f2=f0; mov b1=r0
90 mov r5=0; mov f3=f0; mov b2=r0
91 mov r6=0; mov f4=f0; mov b3=r0
92 mov r7=0; mov f5=f0; mov b4=r0
93 ldf.fill f12=[sp]; mov f13=f0; mov b5=r0
94 ldf.fill f14=[sp]; ldf.fill f15=[sp]; mov f16=f0
95 ldf.fill f17=[sp]; ldf.fill f18=[sp]; mov f19=f0
96 ldf.fill f20=[sp]; ldf.fill f21=[sp]; mov f22=f0
97 ldf.fill f23=[sp]; ldf.fill f24=[sp]; mov f25=f0
98 ldf.fill f26=[sp]; ldf.fill f27=[sp]; mov f28=f0
99 ldf.fill f29=[sp]; ldf.fill f30=[sp]; mov f31=f0
114 mov loc0=rp
115 mov loc1=r16 // save ar.pfs across ia64_clone
117 mov out0=in0
118 mov out1=in1
119 mov out2=in2
120 mov out3=in3
121 mov out4=in4
122 mov out5=in5
126 mov ar.pfs=loc1
127 mov rp=loc0
142 mov loc0=rp
143 mov loc1=r16 // save ar.pfs across ia64_clone
145 mov out0=in0
146 mov out1=in1
147 mov out2=16 // stacksize (compensates for 16-byte scratch area)
148 mov out3=in3
149 mov out4=in4
150 mov out5=in5
154 mov ar.pfs=loc1
155 mov rp=loc0
173 mov r27=IA64_KR(CURRENT_STACK)
190 mov r8=r13 // return pointer to previously running task
191 mov r13=in0 // set "current" pointer
206 mov r25=IA64_GRANULE_SHIFT<<2
211 mov r25=IA64_TR_CURRENT_STACK
240 mov r17=ar.unat // preserve caller's
273 mov.m ar.rsc=0 // put RSE in mode: enforced lazy, little endian, pl 0
277 mov.m r18=ar.fpsr // preserve fpsr
281 mov.m r19=ar.rnat
282 mov r21=b0
286 mov r22=b1
289 mov.m r29=ar.unat
290 mov.m r20=ar.bspstore
291 mov r23=b2
294 mov r24=b3
298 mov r25=b4
299 mov r26=b5
303 mov r21=ar.lc // I-unit
341 mov r21=pr
348 mov ar.rsc=3 // put RSE back into eager mode, pl 0
366 mov ar.rsc=0 // put RSE into enforced lazy mode
408 mov b0=r21
412 mov b1=r22
416 mov b2=r23
418 mov ar.bspstore=r27
419 mov ar.unat=r29 // establish unat holding the NaT bits for r4-r7
420 mov b3=r24
424 mov b4=r25
428 mov b5=r26
432 mov ar.pfs=r16
436 mov ar.lc=r17
440 mov pr=r28,-1
445 mov ar.unat=r18 // restore caller's unat
446 mov ar.rnat=r30 // must restore after bspstore but before rsc!
447 mov ar.fpsr=r19 // restore fpsr
448 mov ar.rsc=3 // put RSE back into eager mode, pl 0
479 mov r10=0
497 mov r3=NR_syscalls - 1
508 mov b6=r20
514 mov r10=0
533 (p6) mov r10=-1
534 (p6) mov r8=r9
559 mov loc0=rp
560 mov loc2=gp
561 mov out0=r5 // arg
564 mov b6 = r14
568 .ret12: mov gp=loc2
569 mov rp=loc0
570 mov ar.pfs=loc1
597 mov r8=0
609 mov r10=r0 // clear error indication in r10
678 (pUStk) mov r21=0 // r21 <- 0
707 mov r16=ar.bsp // M2 get existing backing store pointer
734 mov r22=r0 // A clear r22
747 mov f6=f0 // F clear f6
751 mov f7=f0 // F clear f7
755 (pUStk) mov r17=1 // A
763 mov f8=f0 // F clear f8
767 mov b6=r18 // I0 restore b6
770 mov f9=f0 // F clear f9
778 mov r19=ar.bsp // M2 get new backing store pointer
780 mov f10=f0 // F clear f10
782 mov r22=r0 // A clear r22
786 mov r19=ar.bsp // M2 get new backing store pointer
787 mov f10=f0 // F clear f10
793 mov.m ar.csd=r0 // M2 clear ar.csd
794 mov.m ar.ccv=r0 // M2 clear ar.ccv
795 mov b7=r14 // I0 clear b7 (hint with __kernel_syscall_via_epc)
797 mov.m ar.ssd=r0 // M2 clear ar.ssd
798 mov f11=f0 // F clear f11
822 (pUStk) mov r21=0 // r21 <- 0
871 mov ar.csd=r30
872 mov ar.ssd=r31
879 mov b6=r28
883 mov b7=r29
903 mov ar.ccv=r15
908 (pUStk) mov r18=IA64_KR(CURRENT)// M2 (12 cycle read latency)
959 (pUStk) mov r17=1
974 mov r16=ar.bsp // get existing backing store pointer
980 mov r16=ar.bsp // get existing backing store pointer
993 mov r19=ar.bsp // get new backing store pointer
1021 mov ar.rsc=r19 // load ar.rsc to be used for "loadrs"
1023 mov in1=0
1039 mov loc1=0
1041 mov loc2=0
1043 mov loc3=0
1044 mov loc4=0
1048 mov loc5=0
1052 mov loc6=0
1053 mov loc7=0
1061 mov loc1=0
1062 mov loc2=0
1064 mov loc3=0
1065 mov loc4=0
1066 mov loc5=0
1067 mov loc6=0
1068 mov loc7=0
1071 mov loc8=0
1072 mov loc9=0
1074 mov loc10=0
1075 mov loc11=0
1086 mov ar.unat=r25 // M2
1088 (pLvSys)mov r19=r0 // A clear r19 for leave_syscall, no-op otherwise
1090 (pUStk) mov ar.bspstore=r23 // M2
1092 (pLvSys)mov r16=r0 // A clear r16 for leave_syscall, no-op otherwise
1095 mov ar.pfs=r26 // I0
1096 (pLvSys)mov r17=r0 // A clear r17 for leave_syscall, no-op otherwise
1099 mov b0=r21 // I0
1100 (pLvSys)mov r18=r0 // A clear r18 for leave_syscall, no-op otherwise
1102 mov ar.fpsr=r20 // M2
1106 (pUStk) mov ar.rnat=r24 // M2 must happen with RSE in lazy mode
1108 (pLvSys)mov r2=r0
1110 mov ar.rsc=r27 // M2
1111 mov pr=r31,-1 // I0
1166 (p7) mov r10=-1
1178 mov loc0=rp
1179 mov out0=r8 // Address of previous task
1182 .ret11: mov ar.pfs=loc1
1183 mov rp=loc0
1198 mov r9=ar.unat
1199 mov loc0=rp // save return address
1200 mov out0=0 // there is no "oldset"
1202 (pSys) mov out2=1 // out2==1 => we're in a syscall
1204 (pNonSys) mov out2=0 // out2==0 => not a syscall
1215 mov rp=loc0
1217 mov ar.unat=r9
1218 mov ar.pfs=loc1
1257 mov.sptk b7=r8,ia64_leave_kernel
1259 mov ar.unat=r9
1268 mov r16=r0
1286 mov loc0=rp
1287 mov r16=loc1
1298 mov out1=r13 // current
1302 mov b6=loc2
1303 mov loc2=gp // save gp across indirect function call
1306 mov out1=in1 // arg
1308 1: mov gp=loc2 // restore gp
1316 mov ar.pfs=loc1
1317 mov rp=loc0
1334 mov out3 = r0
1336 mov out2 = b0
1338 mov out1 = r1;
1343 mov b0 = r3
1354 mov loc1 = b0
1355 mov out0 = b0
1356 mov loc2 = r8
1357 mov loc3 = r15
1360 mov out1 = in2
1361 mov b6 = r3
1365 mov ar.pfs = loc0
1366 mov b0 = loc1
1367 mov r8 = loc2
1368 mov r15 = loc3
1385 mov loc1 = b0
1386 mov out0 = b0
1387 mov loc2 = r8
1388 mov loc3 = r15
1391 mov out1 = in2
1392 mov b6 = r3
1396 mov ar.pfs = loc0
1397 mov b0 = loc1
1398 mov r8 = loc2
1399 mov r15 = loc3
1406 mov r3 = b0
1409 mov b6 = r2
1410 mov b7 = r3
1414 mov b0 = r42
1415 mov r1 = r41
1416 mov ar.pfs = r40