443 lines
18 KiB
ArmAsm
443 lines
18 KiB
ArmAsm
/*
|
|
* ARM NEON optimised FFT
|
|
*
|
|
* Copyright (c) 2009 Mans Rullgard <mans@mansr.com>
|
|
* Copyright (c) 2009 Naotoshi Nojiri
|
|
* Copyright (c) 2014 Janne Grunau <janne-libav@jannau.net>
|
|
*
|
|
* This algorithm (though not any of the implementation details) is
|
|
* based on libdjbfft by D. J. Bernstein.
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "libavutil/aarch64/asm.S"
|
|
|
|
#define M_SQRT1_2 0.70710678118654752440
|
|
|
|
.macro transpose d0, d1, s0, s1
|
|
trn1 \d0, \s0, \s1
|
|
trn2 \d1, \s0, \s1
|
|
.endm
|
|
|
|
|
|
function fft4_neon
|
|
ld1 {v0.2s,v1.2s,v2.2s,v3.2s}, [x0]
|
|
|
|
fadd v4.2s, v0.2s, v1.2s // r0+r1,i0+i1
|
|
fsub v6.2s, v0.2s, v1.2s // r0-r1,i0-i1
|
|
|
|
ext v16.8b, v2.8b, v3.8b, #4
|
|
ext v17.8b, v3.8b, v2.8b, #4
|
|
|
|
fadd v5.2s, v2.2s, v3.2s // i2+i3,r2+r3
|
|
fsub v7.2s, v16.2s, v17.2s // r3-r2,i2-i3
|
|
|
|
fadd v0.2s, v4.2s, v5.2s
|
|
fsub v2.2s, v4.2s, v5.2s
|
|
fadd v1.2s, v6.2s, v7.2s
|
|
fsub v3.2s, v6.2s, v7.2s
|
|
|
|
st1 {v0.2s,v1.2s,v2.2s,v3.2s}, [x0]
|
|
|
|
ret
|
|
endfunc
|
|
|
|
function fft8_neon
|
|
mov x1, x0
|
|
ld1 {v0.2s, v1.2s, v2.2s, v3.2s}, [x0], #32
|
|
ld1 {v16.2s,v17.2s,v18.2s,v19.2s}, [x0]
|
|
ext v22.8b, v2.8b, v3.8b, #4
|
|
ext v23.8b, v3.8b, v2.8b, #4
|
|
fadd v4.2s, v16.2s, v17.2s // r4+r5,i4+i5
|
|
fadd v5.2s, v18.2s, v19.2s // r6+r7,i6+i7
|
|
fsub v17.2s, v16.2s, v17.2s // r4-r5,i4-i5
|
|
fsub v19.2s, v18.2s, v19.2s // r6-r7,i6-i7
|
|
rev64 v27.2s, v28.2s // ???
|
|
fadd v20.2s, v0.2s, v1.2s // r0+r1,i0+i1
|
|
fadd v21.2s, v2.2s, v3.2s // r2+r3,i2+i3
|
|
fmul v26.2s, v17.2s, v28.2s // -a2r*w,a2i*w
|
|
ext v6.8b, v4.8b, v5.8b, #4
|
|
ext v7.8b, v5.8b, v4.8b, #4
|
|
fmul v27.2s, v19.2s, v27.2s // a3r*w,-a3i*w
|
|
fsub v23.2s, v22.2s, v23.2s // i2-i3,r3-r2
|
|
fsub v22.2s, v0.2s, v1.2s // r0-r1,i0-i1
|
|
fmul v24.2s, v17.2s, v28.s[1] // a2r*w,a2i*w
|
|
fmul v25.2s, v19.2s, v28.s[1] // a3r*w,a3i*w
|
|
fadd v0.2s, v20.2s, v21.2s
|
|
fsub v2.2s, v20.2s, v21.2s
|
|
fadd v1.2s, v22.2s, v23.2s
|
|
rev64 v26.2s, v26.2s
|
|
rev64 v27.2s, v27.2s
|
|
fsub v3.2s, v22.2s, v23.2s
|
|
fsub v6.2s, v6.2s, v7.2s
|
|
fadd v24.2s, v24.2s, v26.2s // a2r+a2i,a2i-a2r t1,t2
|
|
fadd v25.2s, v25.2s, v27.2s // a3r-a3i,a3i+a3r t5,t6
|
|
fadd v7.2s, v4.2s, v5.2s
|
|
fsub v18.2s, v2.2s, v6.2s
|
|
ext v26.8b, v24.8b, v25.8b, #4
|
|
ext v27.8b, v25.8b, v24.8b, #4
|
|
fadd v2.2s, v2.2s, v6.2s
|
|
fsub v16.2s, v0.2s, v7.2s
|
|
fadd v5.2s, v25.2s, v24.2s
|
|
fsub v4.2s, v26.2s, v27.2s
|
|
fadd v0.2s, v0.2s, v7.2s
|
|
fsub v17.2s, v1.2s, v5.2s
|
|
fsub v19.2s, v3.2s, v4.2s
|
|
fadd v3.2s, v3.2s, v4.2s
|
|
fadd v1.2s, v1.2s, v5.2s
|
|
|
|
st1 {v16.2s,v17.2s,v18.2s,v19.2s}, [x0]
|
|
st1 {v0.2s, v1.2s, v2.2s, v3.2s}, [x1]
|
|
|
|
ret
|
|
endfunc
|
|
|
|
function fft16_neon
|
|
mov x1, x0
|
|
ld1 {v0.2s, v1.2s, v2.2s, v3.2s}, [x0], #32
|
|
ld1 {v16.2s,v17.2s,v18.2s,v19.2s}, [x0], #32
|
|
ext v22.8b, v2.8b, v3.8b, #4
|
|
ext v23.8b, v3.8b, v2.8b, #4
|
|
fadd v4.2s, v16.2s, v17.2s // r4+r5,i4+i5
|
|
fadd v5.2s, v18.2s, v19.2s // r6+r7,i6+i7
|
|
fsub v17.2s, v16.2s, v17.2s // r4-r5,i4-i5
|
|
fsub v19.2s, v18.2s, v19.2s // r6-r7,i6-i7
|
|
rev64 v27.2s, v28.2s // ???
|
|
fadd v20.2s, v0.2s, v1.2s // r0+r1,i0+i1
|
|
fadd v21.2s, v2.2s, v3.2s // r2+r3,i2+i3
|
|
fmul v26.2s, v17.2s, v28.2s // -a2r*w,a2i*w
|
|
ext v6.8b, v4.8b, v5.8b, #4
|
|
ext v7.8b, v5.8b, v4.8b, #4
|
|
fmul v27.2s, v19.2s, v27.2s // a3r*w,-a3i*w
|
|
fsub v23.2s, v22.2s, v23.2s // i2-i3,r3-r2
|
|
fsub v22.2s, v0.2s, v1.2s // r0-r1,i0-i1
|
|
fmul v24.2s, v17.2s, v28.s[1] // a2r*w,a2i*w
|
|
fmul v25.2s, v19.2s, v28.s[1] // a3r*w,a3i*w
|
|
fadd v0.2s, v20.2s, v21.2s
|
|
fsub v2.2s, v20.2s, v21.2s
|
|
fadd v1.2s, v22.2s, v23.2s
|
|
rev64 v26.2s, v26.2s
|
|
rev64 v27.2s, v27.2s
|
|
fsub v3.2s, v22.2s, v23.2s
|
|
fsub v6.2s, v6.2s, v7.2s
|
|
fadd v24.2s, v24.2s, v26.2s // a2r+a2i,a2i-a2r t1,t2
|
|
fadd v25.2s, v25.2s, v27.2s // a3r-a3i,a3i+a3r t5,t6
|
|
fadd v7.2s, v4.2s, v5.2s
|
|
fsub v18.2s, v2.2s, v6.2s
|
|
ld1 {v20.4s,v21.4s}, [x0], #32
|
|
ld1 {v22.4s,v23.4s}, [x0], #32
|
|
ext v26.8b, v24.8b, v25.8b, #4
|
|
ext v27.8b, v25.8b, v24.8b, #4
|
|
fadd v2.2s, v2.2s, v6.2s
|
|
fsub v16.2s, v0.2s, v7.2s
|
|
fadd v5.2s, v25.2s, v24.2s
|
|
fsub v4.2s, v26.2s, v27.2s
|
|
transpose v24.2d, v25.2d, v20.2d, v22.2d
|
|
transpose v26.2d, v27.2d, v21.2d, v23.2d
|
|
fadd v0.2s, v0.2s, v7.2s
|
|
fsub v17.2s, v1.2s, v5.2s
|
|
fsub v19.2s, v3.2s, v4.2s
|
|
fadd v3.2s, v3.2s, v4.2s
|
|
fadd v1.2s, v1.2s, v5.2s
|
|
ext v20.16b, v21.16b, v21.16b, #4
|
|
ext v21.16b, v23.16b, v23.16b, #4
|
|
|
|
zip1 v0.2d, v0.2d, v1.2d // {z[0], z[1]}
|
|
zip1 v1.2d, v2.2d, v3.2d // {z[2], z[3]}
|
|
zip1 v2.2d, v16.2d, v17.2d // {z[o1], z[o1+1]}
|
|
zip1 v3.2d, v18.2d, v19.2d // {z[o1+2],z[o1+3]}
|
|
|
|
// 2 x fft4
|
|
transpose v22.2d, v23.2d, v20.2d, v21.2d
|
|
|
|
fadd v4.4s, v24.4s, v25.4s
|
|
fadd v5.4s, v26.4s, v27.4s
|
|
fsub v6.4s, v24.4s, v25.4s
|
|
fsub v7.4s, v22.4s, v23.4s
|
|
|
|
ld1 {v23.4s}, [x14]
|
|
|
|
fadd v24.4s, v4.4s, v5.4s // {z[o2+0],z[o2+1]}
|
|
fsub v26.4s, v4.4s, v5.4s // {z[o2+2],z[o2+3]}
|
|
fadd v25.4s, v6.4s, v7.4s // {z[o3+0],z[o3+1]}
|
|
fsub v27.4s, v6.4s, v7.4s // {z[o3+2],z[o3+3]}
|
|
|
|
//fft_pass_neon_16
|
|
rev64 v7.4s, v25.4s
|
|
fmul v25.4s, v25.4s, v23.s[1]
|
|
fmul v7.4s, v7.4s, v29.4s
|
|
fmla v25.4s, v7.4s, v23.s[3] // {t1a,t2a,t5a,t6a}
|
|
|
|
zip1 v20.4s, v24.4s, v25.4s
|
|
zip2 v21.4s, v24.4s, v25.4s
|
|
fneg v22.4s, v20.4s
|
|
fadd v4.4s, v21.4s, v20.4s
|
|
fsub v6.4s, v20.4s, v21.4s // just the second half
|
|
fadd v5.4s, v21.4s, v22.4s // just the first half
|
|
|
|
tbl v4.16b, {v4.16b}, v30.16b // trans4_float
|
|
tbl v5.16b, {v5.16b,v6.16b}, v31.16b // trans8_float
|
|
|
|
fsub v20.4s, v0.4s, v4.4s // {z[o2],z[o2+1]}
|
|
fadd v16.4s, v0.4s, v4.4s // {z[0], z[1]}
|
|
fsub v22.4s, v2.4s, v5.4s // {z[o3],z[o3+1]}
|
|
fadd v18.4s, v2.4s, v5.4s // {z[o1],z[o1+1]}
|
|
|
|
//second half
|
|
rev64 v6.4s, v26.4s
|
|
fmul v26.4s, v26.4s, v23.s[2]
|
|
rev64 v7.4s, v27.4s
|
|
fmul v27.4s, v27.4s, v23.s[3]
|
|
fmul v6.4s, v6.4s, v29.4s
|
|
fmul v7.4s, v7.4s, v29.4s
|
|
fmla v26.4s, v6.4s, v23.s[2] // {t1,t2,t5,t6}
|
|
fmla v27.4s, v7.4s, v23.s[1] // {t1a,t2a,t5a,t6a}
|
|
|
|
zip1 v24.4s, v26.4s, v27.4s
|
|
zip2 v25.4s, v26.4s, v27.4s
|
|
fneg v26.4s, v24.4s
|
|
fadd v4.4s, v25.4s, v24.4s
|
|
fsub v6.4s, v24.4s, v25.4s // just the second half
|
|
fadd v5.4s, v25.4s, v26.4s // just the first half
|
|
|
|
tbl v4.16b, {v4.16b}, v30.16b // trans4_float
|
|
tbl v5.16b, {v5.16b,v6.16b}, v31.16b // trans8_float
|
|
|
|
fadd v17.4s, v1.4s, v4.4s // {z[2], z[3]}
|
|
fsub v21.4s, v1.4s, v4.4s // {z[o2+2],z[o2+3]}
|
|
fadd v19.4s, v3.4s, v5.4s // {z[o1+2],z[o1+3]}
|
|
fsub v23.4s, v3.4s, v5.4s // {z[o3+2],z[o3+3]}
|
|
|
|
st1 {v16.4s,v17.4s}, [x1], #32
|
|
st1 {v18.4s,v19.4s}, [x1], #32
|
|
st1 {v20.4s,v21.4s}, [x1], #32
|
|
st1 {v22.4s,v23.4s}, [x1], #32
|
|
|
|
ret
|
|
endfunc
|
|
|
|
|
|
const trans4_float, align=4
|
|
.byte 0, 1, 2, 3
|
|
.byte 8, 9, 10, 11
|
|
.byte 4, 5, 6, 7
|
|
.byte 12, 13, 14, 15
|
|
endconst
|
|
|
|
const trans8_float, align=4
|
|
.byte 24, 25, 26, 27
|
|
.byte 0, 1, 2, 3
|
|
.byte 28, 29, 30, 31
|
|
.byte 4, 5, 6, 7
|
|
endconst
|
|
|
|
function fft_pass_neon
|
|
sub x6, x2, #1 // n - 1, loop counter
|
|
lsl x5, x2, #3 // 2 * n * sizeof FFTSample
|
|
lsl x1, x2, #4 // 2 * n * sizeof FFTComplex
|
|
add x5, x4, x5 // wim
|
|
add x3, x1, x2, lsl #5 // 4 * n * sizeof FFTComplex
|
|
add x2, x0, x2, lsl #5 // &z[o2]
|
|
add x3, x0, x3 // &z[o3]
|
|
add x1, x0, x1 // &z[o1]
|
|
ld1 {v20.4s},[x2] // {z[o2],z[o2+1]}
|
|
ld1 {v22.4s},[x3] // {z[o3],z[o3+1]}
|
|
ld1 {v4.2s}, [x4], #8 // {wre[0],wre[1]}
|
|
trn2 v25.2d, v20.2d, v22.2d
|
|
sub x5, x5, #4 // wim--
|
|
trn1 v24.2d, v20.2d, v22.2d
|
|
ld1 {v5.s}[0], [x5], x7 // d5[0] = wim[-1]
|
|
rev64 v7.4s, v25.4s
|
|
fmul v25.4s, v25.4s, v4.s[1]
|
|
ld1 {v16.4s}, [x0] // {z[0],z[1]}
|
|
fmul v7.4s, v7.4s, v29.4s
|
|
ld1 {v17.4s}, [x1] // {z[o1],z[o1+1]}
|
|
prfm pldl1keep, [x2, #16]
|
|
prfm pldl1keep, [x3, #16]
|
|
fmla v25.4s, v7.4s, v5.s[0] // {t1a,t2a,t5a,t6a}
|
|
prfm pldl1keep, [x0, #16]
|
|
prfm pldl1keep, [x1, #16]
|
|
|
|
zip1 v20.4s, v24.4s, v25.4s
|
|
zip2 v21.4s, v24.4s, v25.4s
|
|
fneg v22.4s, v20.4s
|
|
fadd v4.4s, v21.4s, v20.4s
|
|
fsub v6.4s, v20.4s, v21.4s // just the second half
|
|
fadd v5.4s, v21.4s, v22.4s // just the first half
|
|
|
|
tbl v4.16b, {v4.16b}, v30.16b // trans4_float
|
|
tbl v5.16b, {v5.16b,v6.16b}, v31.16b // trans8_float
|
|
|
|
fadd v20.4s, v16.4s, v4.4s
|
|
fsub v22.4s, v16.4s, v4.4s
|
|
fadd v21.4s, v17.4s, v5.4s
|
|
st1 {v20.4s}, [x0], #16 // {z[0], z[1]}
|
|
fsub v23.4s, v17.4s, v5.4s
|
|
|
|
st1 {v21.4s}, [x1], #16 // {z[o1],z[o1+1]}
|
|
st1 {v22.4s}, [x2], #16 // {z[o2],z[o2+1]}
|
|
st1 {v23.4s}, [x3], #16 // {z[o3],z[o3+1]}
|
|
1:
|
|
ld1 {v20.4s},[x2] // {z[o2],z[o2+1]}
|
|
ld1 {v22.4s},[x3] // {z[o3],z[o3+1]}
|
|
ld1 {v4.2s}, [x4], #8 // {wre[0],wre[1]}
|
|
transpose v26.2d, v27.2d, v20.2d, v22.2d
|
|
ld1 {v5.2s}, [x5], x7 // {wim[-1],wim[0]}
|
|
rev64 v6.4s, v26.4s
|
|
fmul v26.4s, v26.4s, v4.s[0]
|
|
rev64 v7.4s, v27.4s
|
|
fmul v27.4s, v27.4s, v4.s[1]
|
|
fmul v6.4s, v6.4s, v29.4s
|
|
fmul v7.4s, v7.4s, v29.4s
|
|
ld1 {v16.4s},[x0] // {z[0],z[1]}
|
|
fmla v26.4s, v6.4s, v5.s[1] // {t1,t2,t5,t6}
|
|
fmla v27.4s, v7.4s, v5.s[0] // {t1a,t2a,t5a,t6a}
|
|
ld1 {v17.4s},[x1] // {z[o1],z[o1+1]}
|
|
|
|
subs x6, x6, #1 // n--
|
|
|
|
zip1 v20.4s, v26.4s, v27.4s
|
|
zip2 v21.4s, v26.4s, v27.4s
|
|
fneg v22.4s, v20.4s
|
|
fadd v4.4s, v21.4s, v20.4s
|
|
fsub v6.4s, v20.4s, v21.4s // just the second half
|
|
fadd v5.4s, v21.4s, v22.4s // just the first half
|
|
|
|
tbl v4.16b, {v4.16b}, v30.16b // trans4_float
|
|
tbl v5.16b, {v5.16b,v6.16b}, v31.16b // trans8_float
|
|
|
|
fadd v20.4s, v16.4s, v4.4s
|
|
fsub v22.4s, v16.4s, v4.4s
|
|
fadd v21.4s, v17.4s, v5.4s
|
|
st1 {v20.4s}, [x0], #16 // {z[0], z[1]}
|
|
fsub v23.4s, v17.4s, v5.4s
|
|
|
|
st1 {v21.4s}, [x1], #16 // {z[o1],z[o1+1]}
|
|
st1 {v22.4s}, [x2], #16 // {z[o2],z[o2+1]}
|
|
st1 {v23.4s}, [x3], #16 // {z[o3],z[o3+1]}
|
|
b.ne 1b
|
|
|
|
ret
|
|
endfunc
|
|
|
|
.macro def_fft n, n2, n4
|
|
function fft\n\()_neon, align=6
|
|
sub sp, sp, #16
|
|
stp x28, x30, [sp]
|
|
add x28, x0, #\n4*2*8
|
|
bl fft\n2\()_neon
|
|
mov x0, x28
|
|
bl fft\n4\()_neon
|
|
add x0, x28, #\n4*1*8
|
|
bl fft\n4\()_neon
|
|
sub x0, x28, #\n4*2*8
|
|
ldp x28, x30, [sp], #16
|
|
movrel x4, X(ff_cos_\n)
|
|
mov x2, #\n4>>1
|
|
b fft_pass_neon
|
|
endfunc
|
|
.endm
|
|
|
|
def_fft 32, 16, 8
|
|
def_fft 64, 32, 16
|
|
def_fft 128, 64, 32
|
|
def_fft 256, 128, 64
|
|
def_fft 512, 256, 128
|
|
def_fft 1024, 512, 256
|
|
def_fft 2048, 1024, 512
|
|
def_fft 4096, 2048, 1024
|
|
def_fft 8192, 4096, 2048
|
|
def_fft 16384, 8192, 4096
|
|
def_fft 32768, 16384, 8192
|
|
def_fft 65536, 32768, 16384
|
|
|
|
function ff_fft_calc_neon, export=1
|
|
prfm pldl1keep, [x1]
|
|
movrel x10, trans4_float
|
|
ldr w2, [x0]
|
|
movrel x11, trans8_float
|
|
sub w2, w2, #2
|
|
movrel x3, fft_tab_neon
|
|
ld1 {v30.16b}, [x10]
|
|
mov x7, #-8
|
|
movrel x12, pmmp
|
|
ldr x3, [x3, x2, lsl #3]
|
|
movrel x13, mppm
|
|
movrel x14, X(ff_cos_16)
|
|
ld1 {v31.16b}, [x11]
|
|
mov x0, x1
|
|
ld1 {v29.4s}, [x12] // pmmp
|
|
ld1 {v28.4s}, [x13]
|
|
br x3
|
|
endfunc
|
|
|
|
function ff_fft_permute_neon, export=1
|
|
mov x6, #1
|
|
ldr w2, [x0] // nbits
|
|
ldr x3, [x0, #16] // tmp_buf
|
|
ldr x0, [x0, #8] // revtab
|
|
lsl x6, x6, x2
|
|
mov x2, x6
|
|
1:
|
|
ld1 {v0.2s,v1.2s}, [x1], #16
|
|
ldr w4, [x0], #4
|
|
uxth w5, w4
|
|
lsr w4, w4, #16
|
|
add x5, x3, x5, lsl #3
|
|
add x4, x3, x4, lsl #3
|
|
st1 {v0.2s}, [x5]
|
|
st1 {v1.2s}, [x4]
|
|
subs x6, x6, #2
|
|
b.gt 1b
|
|
|
|
sub x1, x1, x2, lsl #3
|
|
1:
|
|
ld1 {v0.4s,v1.4s}, [x3], #32
|
|
st1 {v0.4s,v1.4s}, [x1], #32
|
|
subs x2, x2, #4
|
|
b.gt 1b
|
|
|
|
ret
|
|
endfunc
|
|
|
|
const fft_tab_neon, relocate=1
|
|
.quad fft4_neon
|
|
.quad fft8_neon
|
|
.quad fft16_neon
|
|
.quad fft32_neon
|
|
.quad fft64_neon
|
|
.quad fft128_neon
|
|
.quad fft256_neon
|
|
.quad fft512_neon
|
|
.quad fft1024_neon
|
|
.quad fft2048_neon
|
|
.quad fft4096_neon
|
|
.quad fft8192_neon
|
|
.quad fft16384_neon
|
|
.quad fft32768_neon
|
|
.quad fft65536_neon
|
|
endconst
|
|
|
|
const pmmp, align=4
|
|
.float +1.0, -1.0, -1.0, +1.0
|
|
endconst
|
|
|
|
const mppm, align=4
|
|
.float -M_SQRT1_2, M_SQRT1_2, M_SQRT1_2, -M_SQRT1_2
|
|
endconst
|