Lines Matching +full:1 +full:mib
72 and ptr2 = -(MIN1+1), dest // aligned address
75 } { .mib
80 { .mib
82 } { .mib
83 sub bytecnt = (MIN1+1), tmp // NB: # of bytes to move is 1 higher than loopcnt
88 (p_unalgn) add ptr1 = (MIN1+1), ptr2 // after alignment
92 { .mib
95 } { .mib
99 { .mib
101 (p_unalgn) tbit.nz.unc p_y, p_n = bytecnt, 1 // should we do a st2 ?
102 } { .mib
107 mov tmp = LINE_SIZE+1 // for compare
112 (p_y) st2 [ptr2] = value,-1 //
113 (p_n) add ptr2 = 1, ptr2 //
120 (p_yy) add cnt = -1, cnt //
124 { .mib
134 and cnt = (LINE_SIZE-1), cnt // remainder
136 mov loopcnt = PREF_AHEAD-1 // default prefetch loop
140 (p_scr) add loopcnt = -1, linecnt //
145 add tmp = -1, linecnt // next loop count
149 { .mib
204 and cnt = (LINE_SIZE-1), cnt // remainder
206 mov loopcnt = PREF_AHEAD-1 // default prefetch loop
210 (p_scr) add loopcnt = -1, linecnt
215 add tmp = -1, linecnt // next loop count
219 { .mib
247 { .mib
253 { .mib
257 { .mib
259 add loopcnt = -1, loopcnt
262 { .mib
277 { .mib
298 { .mib
303 { .mib
305 tbit.nz.unc p_yy, p0 = cnt, 1 // should we terminate with a st2 ?
307 { .mib
312 { .mib
316 { .mib
331 (p_y) st1 [ptr1] = value, 1 // fill 1 (odd-aligned) byte [15, 14 (or less) left]
332 (p_y) add cnt = -1, cnt
345 add ptr3 = -1, ptr3 // last store
346 tbit.nz p_scr, p0 = cnt, 1 // will there be a st2 at the end ?