--- trunk/xvidcore/src/image/ia64_asm/interpolate8x8_ia64.s 2002/06/26 15:26:50 243 +++ trunk/xvidcore/src/image/ia64_asm/interpolate8x8_ia64.s 2002/06/26 15:56:35 244 @@ -1,236 +1,236 @@ - - .file "interpolate8x8_ia64.s" - .pred.safe_across_calls p1-p5,p16-p63 - .text - .align 16 - .global interpolate8x8_halfpel_h_ia64# - .proc interpolate8x8_halfpel_h_ia64# -interpolate8x8_halfpel_h_ia64: - LL=3 - SL=1 - SL2=1 - OL=1 - OL2=1 - AVL=1 - AL=1 - STL=3 - - alloc r9=ar.pfs,4, 60,0,64 - - mov r20 = ar.lc - mov r21 = pr - - dep.z r22 = r33,3,3 // rshift of src - - and r14 = -8,r33 // align src - mov r15 = r32 // get dest - mov r16 = r34 // stride - sub r17 = 1,r35 // 1-rounding - ;; - - add r18 = 8,r14 - mux1 r17 = r17, @brcst // broadcast 1-rounding - - sub r24 = 64,r22 // lshift of src - add r26 = 8,r22 // rshift of src+1 - sub r27 = 56,r22 // lshift of src+1 - - mov ar.lc = 7 // loopcounter - mov ar.ec = LL + SL +OL + AVL + AL + STL // sum of latencies - mov pr.rot = 1 << 16 // init pr regs for sw-pipeling - - ;; - .rotr ald1[LL+1],ald2[LL+1],shru1[SL+1],shl1[SL+1],shru2[SL+1],shl2[SL+1],or1[OL+1],or2[OL+1+AL],add1[AL+1],avg[AVL+1] - .rotp aldp[LL], sh1p[SL], or1p[OL], addp[AL], pavg1p[AVL],stp[STL] - - -loop_interpolate: - (aldp[0]) ld8 ald1[0] = [r14],r16 // load aligned src - (aldp[0]) ld8 ald2[0] = [r18],r16 // and aligned src+8 - - (sh1p[0]) shr.u shru1[0] = ald1[LL],r22 // get src - (sh1p[0]) shl shl1[0] = ald2[LL],r27 - (sh1p[0]) shr.u shru2[0] = ald1[LL],r26 // get src+1 - (sh1p[0]) shl shl2[0] = ald2[LL],r24 - - (or1p[0]) or or1[0] = shru1[SL],shl2[SL] // merge things - (or1p[0]) or or2[0] = shru2[SL],shl1[SL] - - (addp[0]) padd1.uus add1[0] = or1[OL],r17 // add 1-rounding - - (pavg1p[0]) pavg1 avg[0] = add1[AL],or2[OL+AL] // parallel average - - (stp[0]) st8 [r15] = avg[AVL] // store results - (stp[0]) add r15 = r15,r16 - - - - - br.ctop.sptk.few loop_interpolate - ;; - mov ar.lc = r20 - mov pr = r21,-1 - br.ret.sptk.many b0 - .endp interpolate8x8_halfpel_h_ia64# - - .align 16 - .global interpolate8x8_halfpel_v_ia64# - .proc interpolate8x8_halfpel_v_ia64# -interpolate8x8_halfpel_v_ia64: - LL=3 - SL=1 - SL2=1 - OL=1 - OL2=1 - AVL=1 - AL=1 - STL=3 - - alloc r9=ar.pfs,4, 60,0,64 - - mov r20 = ar.lc - mov r21 = pr - - dep.z r22 = r33,3,3 - - and r14 = -8,r33 - mov r15 = r32 - mov r16 = r34 - sub r17 = 1,r35 - ;; - - add r18 = 8,r14 - add r19 = r14,r16 // src + stride - mux1 r17 = r17, @brcst - - sub r24 = 64,r22 - ;; - add r26 = 8,r19 // src + stride + 8 - - mov ar.lc = 7 - mov ar.ec = LL + SL +OL + AVL + AL + STL - mov pr.rot = 1 << 16 - - ;; - .rotr ald1[LL+1],ald2[LL+1],ald3[LL+1],ald4[LL+1],shru1[SL+1],shl1[SL+1],shru2[SL+1],shl2[SL+1],or1[OL+1],or2[OL+1+AL],add1[AL+1],avg[AVL+1] - .rotp aldp[LL], sh1p[SL], or1p[OL], addp[AL], pavg1p[AVL],stp[STL] - - -loop_interpolate2: - (aldp[0]) ld8 ald1[0] = [r14],r16 - (aldp[0]) ld8 ald2[0] = [r18],r16 - (aldp[0]) ld8 ald3[0] = [r19],r16 - (aldp[0]) ld8 ald4[0] = [r26],r16 - - (sh1p[0]) shr.u shru1[0] = ald1[LL],r22 - (sh1p[0]) shl shl1[0] = ald2[LL],r24 - (sh1p[0]) shr.u shru2[0] = ald3[LL],r22 - (sh1p[0]) shl shl2[0] = ald4[LL],r24 - - (or1p[0]) or or1[0] = shru1[SL],shl1[SL] - (or1p[0]) or or2[0] = shru2[SL],shl2[SL] - - (addp[0]) padd1.uus add1[0] = or1[OL],r17 - - (pavg1p[0]) pavg1 avg[0] = add1[AL],or2[OL+AL] - - (stp[0]) st8 [r15] = avg[AVL] - (stp[0]) add r15 = r15,r16 - - - - - br.ctop.sptk.few loop_interpolate2 - ;; - mov ar.lc = r20 - mov pr = r21,-1 - br.ret.sptk.many b0 - .endp interpolate8x8_halfpel_v_ia64# - - .align 16 - .global interpolate8x8_halfpel_hv_ia64# - .proc interpolate8x8_halfpel_hv_ia64# -interpolate8x8_halfpel_hv_ia64: - LL=3 - SL=1 - SL2=1 - OL=1 - OL2=1 - AVL=1 - AL=1 - STL=3 - - alloc r9=ar.pfs,4, 60,0,64 - - mov r20 = ar.lc - mov r21 = pr - - dep.z r22 = r33,3,3 - - and r14 = -8,r33 - mov r15 = r32 - mov r16 = r34 - sub r17 = 1,r35 - ;; - - add r18 = 8,r14 - add r19 = r14,r16 - mux1 r17 = r17, @brcst - - add r27 = 8,r22 - sub r28 = 56,r22 - sub r24 = 64,r22 - ;; - add r26 = 8,r19 - - mov ar.lc = 7 - mov ar.ec = LL + SL +OL + 2*AVL + AL + STL - mov pr.rot = 1 << 16 - - ;; - .rotr ald1[LL+1],ald2[LL+1],ald3[LL+1],ald4[LL+1],shru1[SL+1],shl1[SL+1],shru2[SL+1],shl2[SL+1],shl3[SL+1],shru3[SL+1],shl4[SL+1],shru4[SL+1],or1[OL+1],or2[OL+1+AL],or3[OL+AL+1],or4[OL+AL+1],add1[AL+1],avg[AVL+1],avg1[AVL+1],avg2[AVL+1] - .rotp aldp[LL], sh1p[SL], or1p[OL], addp[AL],pavg1p[AVL],pavg2p[AVL],stp[STL] - - -loop_interpolate3: - (aldp[0]) ld8 ald1[0] = [r14],r16 - (aldp[0]) ld8 ald2[0] = [r18],r16 - (aldp[0]) ld8 ald3[0] = [r19],r16 - (aldp[0]) ld8 ald4[0] = [r26],r16 - - (sh1p[0]) shr.u shru1[0] = ald1[LL],r22 - (sh1p[0]) shl shl1[0] = ald2[LL],r24 - (sh1p[0]) shr.u shru2[0] = ald3[LL],r22 - (sh1p[0]) shl shl2[0] = ald4[LL],r24 - (sh1p[0]) shr.u shru3[0] = ald1[LL],r27 - (sh1p[0]) shl shl3[0] = ald2[LL],r28 - (sh1p[0]) shr.u shru4[0] = ald3[LL],r27 - (sh1p[0]) shl shl4[0] = ald4[LL],r28 - - - (or1p[0]) or or1[0] = shru1[SL],shl1[SL] - (or1p[0]) or or2[0] = shru2[SL],shl2[SL] - (or1p[0]) or or3[0] = shru3[SL],shl3[SL] - (or1p[0]) or or4[0] = shru4[SL],shl4[SL] - - (addp[0]) padd1.uus add1[0] = or1[OL],r17 - - (pavg1p[0]) pavg1 avg[0] = add1[AL],or2[OL+AL] - (pavg1p[0]) pavg1 avg1[0] = or3[OL+AL],or4[OL+AL] - - (pavg2p[0]) pavg1 avg2[0] = avg[AVL],avg1[AVL] - - (stp[0]) st8 [r15] = avg2[AVL] - (stp[0]) add r15 = r15,r16 - - - - - br.ctop.sptk.few loop_interpolate3 - ;; - mov ar.lc = r20 - mov pr = r21,-1 - br.ret.sptk.many b0 - .endp interpolate8x8_halfpel_hv_ia64# - - + + .file "interpolate8x8_ia64.s" + .pred.safe_across_calls p1-p5,p16-p63 + .text + .align 16 + .global interpolate8x8_halfpel_h_ia64# + .proc interpolate8x8_halfpel_h_ia64# +interpolate8x8_halfpel_h_ia64: + LL=3 + SL=1 + SL2=1 + OL=1 + OL2=1 + AVL=1 + AL=1 + STL=3 + + alloc r9=ar.pfs,4, 60,0,64 + + mov r20 = ar.lc + mov r21 = pr + + dep.z r22 = r33,3,3 // rshift of src + + and r14 = -8,r33 // align src + mov r15 = r32 // get dest + mov r16 = r34 // stride + sub r17 = 1,r35 // 1-rounding + ;; + + add r18 = 8,r14 + mux1 r17 = r17, @brcst // broadcast 1-rounding + + sub r24 = 64,r22 // lshift of src + add r26 = 8,r22 // rshift of src+1 + sub r27 = 56,r22 // lshift of src+1 + + mov ar.lc = 7 // loopcounter + mov ar.ec = LL + SL +OL + AVL + AL + STL // sum of latencies + mov pr.rot = 1 << 16 // init pr regs for sw-pipeling + + ;; + .rotr ald1[LL+1],ald2[LL+1],shru1[SL+1],shl1[SL+1],shru2[SL+1],shl2[SL+1],or1[OL+1],or2[OL+1+AL],add1[AL+1],avg[AVL+1] + .rotp aldp[LL], sh1p[SL], or1p[OL], addp[AL], pavg1p[AVL],stp[STL] + + +loop_interpolate: + (aldp[0]) ld8 ald1[0] = [r14],r16 // load aligned src + (aldp[0]) ld8 ald2[0] = [r18],r16 // and aligned src+8 + + (sh1p[0]) shr.u shru1[0] = ald1[LL],r22 // get src + (sh1p[0]) shl shl1[0] = ald2[LL],r27 + (sh1p[0]) shr.u shru2[0] = ald1[LL],r26 // get src+1 + (sh1p[0]) shl shl2[0] = ald2[LL],r24 + + (or1p[0]) or or1[0] = shru1[SL],shl2[SL] // merge things + (or1p[0]) or or2[0] = shru2[SL],shl1[SL] + + (addp[0]) padd1.uus add1[0] = or1[OL],r17 // add 1-rounding + + (pavg1p[0]) pavg1 avg[0] = add1[AL],or2[OL+AL] // parallel average + + (stp[0]) st8 [r15] = avg[AVL] // store results + (stp[0]) add r15 = r15,r16 + + + + + br.ctop.sptk.few loop_interpolate + ;; + mov ar.lc = r20 + mov pr = r21,-1 + br.ret.sptk.many b0 + .endp interpolate8x8_halfpel_h_ia64# + + .align 16 + .global interpolate8x8_halfpel_v_ia64# + .proc interpolate8x8_halfpel_v_ia64# +interpolate8x8_halfpel_v_ia64: + LL=3 + SL=1 + SL2=1 + OL=1 + OL2=1 + AVL=1 + AL=1 + STL=3 + + alloc r9=ar.pfs,4, 60,0,64 + + mov r20 = ar.lc + mov r21 = pr + + dep.z r22 = r33,3,3 + + and r14 = -8,r33 + mov r15 = r32 + mov r16 = r34 + sub r17 = 1,r35 + ;; + + add r18 = 8,r14 + add r19 = r14,r16 // src + stride + mux1 r17 = r17, @brcst + + sub r24 = 64,r22 + ;; + add r26 = 8,r19 // src + stride + 8 + + mov ar.lc = 7 + mov ar.ec = LL + SL +OL + AVL + AL + STL + mov pr.rot = 1 << 16 + + ;; + .rotr ald1[LL+1],ald2[LL+1],ald3[LL+1],ald4[LL+1],shru1[SL+1],shl1[SL+1],shru2[SL+1],shl2[SL+1],or1[OL+1],or2[OL+1+AL],add1[AL+1],avg[AVL+1] + .rotp aldp[LL], sh1p[SL], or1p[OL], addp[AL], pavg1p[AVL],stp[STL] + + +loop_interpolate2: + (aldp[0]) ld8 ald1[0] = [r14],r16 + (aldp[0]) ld8 ald2[0] = [r18],r16 + (aldp[0]) ld8 ald3[0] = [r19],r16 + (aldp[0]) ld8 ald4[0] = [r26],r16 + + (sh1p[0]) shr.u shru1[0] = ald1[LL],r22 + (sh1p[0]) shl shl1[0] = ald2[LL],r24 + (sh1p[0]) shr.u shru2[0] = ald3[LL],r22 + (sh1p[0]) shl shl2[0] = ald4[LL],r24 + + (or1p[0]) or or1[0] = shru1[SL],shl1[SL] + (or1p[0]) or or2[0] = shru2[SL],shl2[SL] + + (addp[0]) padd1.uus add1[0] = or1[OL],r17 + + (pavg1p[0]) pavg1 avg[0] = add1[AL],or2[OL+AL] + + (stp[0]) st8 [r15] = avg[AVL] + (stp[0]) add r15 = r15,r16 + + + + + br.ctop.sptk.few loop_interpolate2 + ;; + mov ar.lc = r20 + mov pr = r21,-1 + br.ret.sptk.many b0 + .endp interpolate8x8_halfpel_v_ia64# + + .align 16 + .global interpolate8x8_halfpel_hv_ia64# + .proc interpolate8x8_halfpel_hv_ia64# +interpolate8x8_halfpel_hv_ia64: + LL=3 + SL=1 + SL2=1 + OL=1 + OL2=1 + AVL=1 + AL=1 + STL=3 + + alloc r9=ar.pfs,4, 60,0,64 + + mov r20 = ar.lc + mov r21 = pr + + dep.z r22 = r33,3,3 + + and r14 = -8,r33 + mov r15 = r32 + mov r16 = r34 + sub r17 = 1,r35 + ;; + + add r18 = 8,r14 + add r19 = r14,r16 + mux1 r17 = r17, @brcst + + add r27 = 8,r22 + sub r28 = 56,r22 + sub r24 = 64,r22 + ;; + add r26 = 8,r19 + + mov ar.lc = 7 + mov ar.ec = LL + SL +OL + 2*AVL + AL + STL + mov pr.rot = 1 << 16 + + ;; + .rotr ald1[LL+1],ald2[LL+1],ald3[LL+1],ald4[LL+1],shru1[SL+1],shl1[SL+1],shru2[SL+1],shl2[SL+1],shl3[SL+1],shru3[SL+1],shl4[SL+1],shru4[SL+1],or1[OL+1],or2[OL+1+AL],or3[OL+AL+1],or4[OL+AL+1],add1[AL+1],avg[AVL+1],avg1[AVL+1],avg2[AVL+1] + .rotp aldp[LL], sh1p[SL], or1p[OL], addp[AL],pavg1p[AVL],pavg2p[AVL],stp[STL] + + +loop_interpolate3: + (aldp[0]) ld8 ald1[0] = [r14],r16 + (aldp[0]) ld8 ald2[0] = [r18],r16 + (aldp[0]) ld8 ald3[0] = [r19],r16 + (aldp[0]) ld8 ald4[0] = [r26],r16 + + (sh1p[0]) shr.u shru1[0] = ald1[LL],r22 + (sh1p[0]) shl shl1[0] = ald2[LL],r24 + (sh1p[0]) shr.u shru2[0] = ald3[LL],r22 + (sh1p[0]) shl shl2[0] = ald4[LL],r24 + (sh1p[0]) shr.u shru3[0] = ald1[LL],r27 + (sh1p[0]) shl shl3[0] = ald2[LL],r28 + (sh1p[0]) shr.u shru4[0] = ald3[LL],r27 + (sh1p[0]) shl shl4[0] = ald4[LL],r28 + + + (or1p[0]) or or1[0] = shru1[SL],shl1[SL] + (or1p[0]) or or2[0] = shru2[SL],shl2[SL] + (or1p[0]) or or3[0] = shru3[SL],shl3[SL] + (or1p[0]) or or4[0] = shru4[SL],shl4[SL] + + (addp[0]) padd1.uus add1[0] = or1[OL],r17 + + (pavg1p[0]) pavg1 avg[0] = add1[AL],or2[OL+AL] + (pavg1p[0]) pavg1 avg1[0] = or3[OL+AL],or4[OL+AL] + + (pavg2p[0]) pavg1 avg2[0] = avg[AVL],avg1[AVL] + + (stp[0]) st8 [r15] = avg2[AVL] + (stp[0]) add r15 = r15,r16 + + + + + br.ctop.sptk.few loop_interpolate3 + ;; + mov ar.lc = r20 + mov pr = r21,-1 + br.ret.sptk.many b0 + .endp interpolate8x8_halfpel_hv_ia64# + +