233 lines
7.3 KiB
ArmAsm
Executable file
233 lines
7.3 KiB
ArmAsm
Executable file
/*
|
|
* Copyright (c) 2014 Janne Grunau <janne-libav@jannau.net>
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "libavutil/aarch64/asm.S"
|
|
#include "asm-offsets.h"
|
|
|
|
.macro resample_one fmt, es=2
|
|
.ifnc \fmt, dbl
|
|
.macro M_MUL2 x:vararg
|
|
.endm
|
|
.macro M_MLA2 x:vararg
|
|
.endm
|
|
.endif
|
|
function ff_resample_one_\fmt\()_neon, export=1
|
|
sxtw x2, w2
|
|
ldr x9, [x0, #FILTER_BANK]
|
|
ldr w6, [x0, #FILTER_LENGTH]
|
|
ldp w7, w8, [x0, #PHASE_SHIFT] // and phase_mask
|
|
lsr x10, x4, x7 // sample_index
|
|
and x4, x4, x8
|
|
lsl x11, x6, #\es // filter_length * elem_size
|
|
add x3, x3, x10, lsl #\es // src[sample_index]
|
|
madd x9, x11, x4, x9 // filter
|
|
cmp w6, #16
|
|
b.lt 5f
|
|
8: // remaining filter_length at least 16
|
|
subs w6, w6, #16
|
|
LOAD8 v4, v5, v6, v7, x3
|
|
LOAD8 v16, v17, v18, v19, x9
|
|
M_MUL v0, v4, v16, v1
|
|
M_MUL2 v1, v6, v18
|
|
7:
|
|
LOAD8 v20, v21, v22, v23, x3
|
|
M_MLA v0, v5, v17, v1
|
|
M_MLA2 v1, v7, v19
|
|
LOAD8 v24, v25, v26, v27, x9
|
|
M_MLA v0, v20, v24, v1
|
|
M_MLA2 v1, v22, v26
|
|
b.eq 6f
|
|
cmp w6, #16
|
|
M_MLA v0, v21, v25, v1
|
|
M_MLA2 v1, v23, v27
|
|
b.lt 4f
|
|
subs w6, w6, #16
|
|
LOAD8 v4, v5, v6, v7, x3
|
|
LOAD8 v16, v17, v18, v19, x9
|
|
M_MLA v0, v4, v16, v1
|
|
M_MLA2 v1, v6, v18
|
|
b 7b
|
|
6:
|
|
M_MLA v0, v21, v25, v1
|
|
M_MLA2 v1, v23, v27
|
|
STORE_ONE 0, x1, x2, v1
|
|
ret
|
|
5:
|
|
movi v0.16b, #0
|
|
movi v1.16b, #0
|
|
4: // remaining filter_length 1-15
|
|
cmp w6, #4
|
|
b.lt 2f
|
|
subs w6, w6, #4
|
|
LOAD4 v4, v5, x3
|
|
LOAD4 v6, v7, x9
|
|
M_MLA v0, v4, v6, v1
|
|
M_MLA2 v1, v5, v7
|
|
b.eq 0f
|
|
b 4b
|
|
2: // remaining filter_length 1-3
|
|
cmp w6, #2
|
|
b.lt 1f
|
|
LOAD2 2, x3
|
|
LOAD2 3, x9
|
|
subs w6, w6, #2
|
|
M_MLA v0, v2, v3
|
|
b.eq 0f
|
|
1: // remaining filter_length 1
|
|
LOAD1 6, x3
|
|
LOAD1 7, x9
|
|
M_MLA v0, v6, v7
|
|
0:
|
|
STORE_ONE 0, x1, x2, v1
|
|
ret
|
|
endfunc
|
|
|
|
.purgem LOAD1
|
|
.purgem LOAD2
|
|
.purgem LOAD4
|
|
.purgem LOAD8
|
|
.purgem M_MLA
|
|
.purgem M_MLA2
|
|
.purgem M_MUL
|
|
.purgem M_MUL2
|
|
.purgem STORE_ONE
|
|
.endm
|
|
|
|
|
|
.macro LOAD1 d1, addr
|
|
ldr d\d1, [\addr], #8
|
|
.endm
|
|
.macro LOAD2 d1, addr
|
|
ld1 {v\d1\().2d}, [\addr], #16
|
|
.endm
|
|
.macro LOAD4 d1, d2, addr
|
|
ld1 {\d1\().2d,\d2\().2d}, [\addr], #32
|
|
.endm
|
|
.macro LOAD8 d1, d2, d3, d4, addr
|
|
ld1 {\d1\().2d,\d2\().2d,\d3\().2d,\d4\().2d}, [\addr], #64
|
|
.endm
|
|
.macro M_MLA d, r0, r1, d2:vararg
|
|
fmla \d\().2d, \r0\().2d, \r1\().2d
|
|
.endm
|
|
.macro M_MLA2 second:vararg
|
|
M_MLA \second
|
|
.endm
|
|
.macro M_MUL d, r0, r1, d2:vararg
|
|
fmul \d\().2d, \r0\().2d, \r1\().2d
|
|
.endm
|
|
.macro M_MUL2 second:vararg
|
|
M_MUL \second
|
|
.endm
|
|
.macro STORE_ONE rn, addr, idx, d2
|
|
fadd v\rn\().2d, v\rn\().2d, \d2\().2d
|
|
faddp d\rn\(), v\rn\().2d
|
|
str d\rn\(), [\addr, \idx, lsl #3]
|
|
.endm
|
|
|
|
resample_one dbl, 3
|
|
|
|
|
|
.macro LOAD1 d1, addr
|
|
ldr s\d1, [\addr], #4
|
|
.endm
|
|
.macro LOAD2 d1, addr
|
|
ld1 {v\d1\().2s}, [\addr], #8
|
|
.endm
|
|
.macro LOAD4 d1, d2, addr
|
|
ld1 {\d1\().4s}, [\addr], #16
|
|
.endm
|
|
.macro LOAD8 d1, d2, d3, d4, addr
|
|
ld1 {\d1\().4s,\d2\().4s}, [\addr], #32
|
|
.endm
|
|
.macro M_MLA d, r0, r1, d2:vararg
|
|
fmla \d\().4s, \r0\().4s, \r1\().4s
|
|
.endm
|
|
.macro M_MUL d, r0, r1, d2:vararg
|
|
fmul \d\().4s, \r0\().4s, \r1\().4s
|
|
.endm
|
|
.macro STORE_ONE rn, addr, idx, d2
|
|
faddp v\rn\().4s, v\rn\().4s, v\rn\().4s
|
|
faddp s\rn\(), v\rn\().2s
|
|
str s\rn\(), [\addr, \idx, lsl #2]
|
|
.endm
|
|
|
|
resample_one flt
|
|
|
|
|
|
.macro LOAD1 d1, addr
|
|
ldr h\d1, [\addr], #2
|
|
.endm
|
|
.macro LOAD2 d1, addr
|
|
ldr s\d1, [\addr], #4
|
|
.endm
|
|
.macro LOAD4 d1, d2, addr
|
|
ld1 {\d1\().4h}, [\addr], #8
|
|
.endm
|
|
.macro LOAD8 d1, d2, d3, d4, addr
|
|
ld1 {\d1\().4h,\d2\().4h}, [\addr], #16
|
|
.endm
|
|
.macro M_MLA d, r0, r1, d2:vararg
|
|
smlal \d\().4s, \r0\().4h, \r1\().4h
|
|
.endm
|
|
.macro M_MUL d, r0, r1, d2:vararg
|
|
smull \d\().4s, \r0\().4h, \r1\().4h
|
|
.endm
|
|
.macro STORE_ONE rn, addr, idx, d2
|
|
addp v\rn\().4s, v\rn\().4s, v\rn\().4s
|
|
addp v\rn\().4s, v\rn\().4s, v\rn\().4s
|
|
sqrshrn v\rn\().4h, v\rn\().4s, #15
|
|
str h\rn\(), [\addr, \idx, lsl #1]
|
|
.endm
|
|
|
|
resample_one s16, 1
|
|
|
|
|
|
.macro LOAD1 d1, addr
|
|
ldr s\d1, [\addr], #4
|
|
.endm
|
|
.macro LOAD2 d1, addr
|
|
ld1 {v\d1\().2s}, [\addr], #8
|
|
.endm
|
|
.macro LOAD4 d1, d2, addr
|
|
ld1 {\d1\().4s}, [\addr], #16
|
|
.endm
|
|
.macro LOAD8 d1, d2, d3, d4, addr
|
|
ld1 {\d1\().4s,\d2\().4s}, [\addr], #32
|
|
.endm
|
|
.macro M_MLA d1, r0, r1, d2:vararg
|
|
smlal \d1\().2d, \r0\().2s, \r1\().2s
|
|
.ifnb \d2
|
|
smlal2 \d2\().2d, \r0\().4s, \r1\().4s
|
|
.endif
|
|
.endm
|
|
.macro M_MUL d1, r0, r1, d2:vararg
|
|
smull \d1\().2d, \r0\().2s, \r1\().2s
|
|
.ifnb \d2
|
|
smull2 \d2\().2d, \r0\().4s, \r1\().4s
|
|
.endif
|
|
.endm
|
|
.macro STORE_ONE rn, addr, idx, d2
|
|
add v\rn\().2d, v\rn\().2d, \d2\().2d
|
|
addp d\rn\(), v\rn\().2d
|
|
sqrshrn v\rn\().2s, v\rn\().2d, #30
|
|
str s\rn\(), [\addr, \idx, lsl #2]
|
|
.endm
|
|
|
|
resample_one s32
|