Spaces:
Runtime error
Runtime error
;****************************************************************************** | |
;* Copyright (c) 2010 David Conrad | |
;* | |
;* This file is part of FFmpeg. | |
;* | |
;* FFmpeg is free software; you can redistribute it and/or | |
;* modify it under the terms of the GNU Lesser General Public | |
;* License as published by the Free Software Foundation; either | |
;* version 2.1 of the License, or (at your option) any later version. | |
;* | |
;* FFmpeg is distributed in the hope that it will be useful, | |
;* but WITHOUT ANY WARRANTY; without even the implied warranty of | |
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU | |
;* Lesser General Public License for more details. | |
;* | |
;* You should have received a copy of the GNU Lesser General Public | |
;* License along with FFmpeg; if not, write to the Free Software | |
;* 51, Inc., Foundation Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA | |
;****************************************************************************** | |
%include "libavutil/x86/x86util.asm" | |
SECTION_RODATA | |
pw_7: times 8 dw 7 | |
convert_to_unsigned_10bit: times 4 dd 0x200 | |
clip_10bit: times 8 dw 0x3ff | |
cextern pw_3 | |
cextern pw_16 | |
cextern pw_32 | |
cextern pb_80 | |
SECTION .text | |
%macro UNPACK_ADD 6 | |
mov%5 %1, %3 | |
mov%6 m5, %4 | |
mova m4, %1 | |
mova %2, m5 | |
punpcklbw %1, m7 | |
punpcklbw m5, m7 | |
punpckhbw m4, m7 | |
punpckhbw %2, m7 | |
paddw %1, m5 | |
paddw %2, m4 | |
%endmacro | |
%macro HPEL_FILTER 1 | |
; dirac_hpel_filter_v_sse2(uint8_t *dst, uint8_t *src, int stride, int width); | |
cglobal dirac_hpel_filter_v_%1, 4,6,8, dst, src, stride, width, src0, stridex3 | |
mov src0q, srcq | |
lea stridex3q, [3*strideq] | |
sub src0q, stridex3q | |
pxor m7, m7 | |
.loop: | |
; 7*(src[0] + src[1]) | |
UNPACK_ADD m0, m1, [srcq], [srcq + strideq], a,a | |
pmullw m0, [pw_7] | |
pmullw m1, [pw_7] | |
; 3*( ... + src[-2] + src[3]) | |
UNPACK_ADD m2, m3, [src0q + strideq], [srcq + stridex3q], a,a | |
paddw m0, m2 | |
paddw m1, m3 | |
pmullw m0, [pw_3] | |
pmullw m1, [pw_3] | |
; ... - 7*(src[-1] + src[2]) | |
UNPACK_ADD m2, m3, [src0q + strideq*2], [srcq + strideq*2], a,a | |
pmullw m2, [pw_7] | |
pmullw m3, [pw_7] | |
psubw m0, m2 | |
psubw m1, m3 | |
; ... - (src[-3] + src[4]) | |
UNPACK_ADD m2, m3, [src0q], [srcq + strideq*4], a,a | |
psubw m0, m2 | |
psubw m1, m3 | |
paddw m0, [pw_16] | |
paddw m1, [pw_16] | |
psraw m0, 5 | |
psraw m1, 5 | |
packuswb m0, m1 | |
mova [dstq], m0 | |
add dstq, mmsize | |
add srcq, mmsize | |
add src0q, mmsize | |
sub widthd, mmsize | |
jg .loop | |
RET | |
; dirac_hpel_filter_h_sse2(uint8_t *dst, uint8_t *src, int width); | |
cglobal dirac_hpel_filter_h_%1, 3,3,8, dst, src, width | |
dec widthd | |
pxor m7, m7 | |
and widthd, ~(mmsize-1) | |
.loop: | |
; 7*(src[0] + src[1]) | |
UNPACK_ADD m0, m1, [srcq + widthq], [srcq + widthq + 1], u,u | |
pmullw m0, [pw_7] | |
pmullw m1, [pw_7] | |
; 3*( ... + src[-2] + src[3]) | |
UNPACK_ADD m2, m3, [srcq + widthq - 2], [srcq + widthq + 3], u,u | |
paddw m0, m2 | |
paddw m1, m3 | |
pmullw m0, [pw_3] | |
pmullw m1, [pw_3] | |
; ... - 7*(src[-1] + src[2]) | |
UNPACK_ADD m2, m3, [srcq + widthq - 1], [srcq + widthq + 2], u,u | |
pmullw m2, [pw_7] | |
pmullw m3, [pw_7] | |
psubw m0, m2 | |
psubw m1, m3 | |
; ... - (src[-3] + src[4]) | |
UNPACK_ADD m2, m3, [srcq + widthq - 3], [srcq + widthq + 4], u,u | |
psubw m0, m2 | |
psubw m1, m3 | |
paddw m0, [pw_16] | |
paddw m1, [pw_16] | |
psraw m0, 5 | |
psraw m1, 5 | |
packuswb m0, m1 | |
mova [dstq + widthq], m0 | |
sub widthd, mmsize | |
jge .loop | |
RET | |
%endmacro | |
%macro PUT_RECT 1 | |
; void put_rect_clamped(uint8_t *dst, int dst_stride, int16_t *src, int src_stride, int width, int height) | |
cglobal put_signed_rect_clamped_%1, 5,9,3, dst, dst_stride, src, src_stride, w, dst2, src2 | |
mova m0, [pb_80] | |
add wd, (mmsize-1) | |
and wd, ~(mmsize-1) | |
%if ARCH_X86_64 | |
movsxd dst_strideq, dst_strided | |
movsxd src_strideq, src_strided | |
mov r7d, r5m | |
mov r8d, wd | |
%define wspill r8d | |
%define hd r7d | |
%else | |
mov r4m, wd | |
%define wspill r4m | |
%define hd r5mp | |
%endif | |
.loopy: | |
lea src2q, [srcq+src_strideq] | |
lea dst2q, [dstq+dst_strideq] | |
.loopx: | |
sub wd, mmsize | |
mova m1, [srcq +2*wq] | |
mova m2, [src2q+2*wq] | |
packsswb m1, [srcq +2*wq+mmsize] | |
packsswb m2, [src2q+2*wq+mmsize] | |
paddb m1, m0 | |
paddb m2, m0 | |
mova [dstq +wq], m1 | |
mova [dst2q+wq], m2 | |
jg .loopx | |
lea srcq, [srcq+src_strideq*2] | |
lea dstq, [dstq+dst_strideq*2] | |
sub hd, 2 | |
mov wd, wspill | |
jg .loopy | |
RET | |
%endm | |
%macro ADD_RECT 1 | |
; void add_rect_clamped(uint8_t *dst, uint16_t *src, int stride, int16_t *idwt, int idwt_stride, int width, int height) | |
cglobal add_rect_clamped_%1, 7,9,3, dst, src, stride, idwt, idwt_stride, w, h | |
mova m0, [pw_32] | |
add wd, (mmsize-1) | |
and wd, ~(mmsize-1) | |
%if ARCH_X86_64 | |
movsxd strideq, strided | |
movsxd idwt_strideq, idwt_strided | |
mov r8d, wd | |
%define wspill r8d | |
%else | |
mov r5m, wd | |
%define wspill r5m | |
%endif | |
.loop: | |
sub wd, mmsize | |
movu m1, [srcq +2*wq] ; FIXME: ensure alignment | |
paddw m1, m0 | |
psraw m1, 6 | |
movu m2, [srcq +2*wq+mmsize] ; FIXME: ensure alignment | |
paddw m2, m0 | |
psraw m2, 6 | |
paddw m1, [idwtq+2*wq] | |
paddw m2, [idwtq+2*wq+mmsize] | |
packuswb m1, m2 | |
mova [dstq +wq], m1 | |
jg .loop | |
lea srcq, [srcq + 2*strideq] | |
add dstq, strideq | |
lea idwtq, [idwtq+ 2*idwt_strideq] | |
sub hd, 1 | |
mov wd, wspill | |
jg .loop | |
RET | |
%endm | |
%macro ADD_OBMC 2 | |
; void add_obmc(uint16_t *dst, uint8_t *src, int stride, uint8_t *obmc_weight, int yblen) | |
cglobal add_dirac_obmc%1_%2, 6,6,5, dst, src, stride, obmc, yblen | |
pxor m4, m4 | |
.loop: | |
%assign i 0 | |
%rep %1 / mmsize | |
mova m0, [srcq+i] | |
mova m1, m0 | |
punpcklbw m0, m4 | |
punpckhbw m1, m4 | |
mova m2, [obmcq+i] | |
mova m3, m2 | |
punpcklbw m2, m4 | |
punpckhbw m3, m4 | |
pmullw m0, m2 | |
pmullw m1, m3 | |
movu m2, [dstq+2*i] | |
movu m3, [dstq+2*i+mmsize] | |
paddw m0, m2 | |
paddw m1, m3 | |
movu [dstq+2*i], m0 | |
movu [dstq+2*i+mmsize], m1 | |
%assign i i+mmsize | |
%endrep | |
lea srcq, [srcq+strideq] | |
lea dstq, [dstq+2*strideq] | |
add obmcq, 32 | |
sub yblend, 1 | |
jg .loop | |
RET | |
%endm | |
INIT_MMX | |
ADD_OBMC 8, mmx | |
INIT_XMM | |
PUT_RECT sse2 | |
ADD_RECT sse2 | |
HPEL_FILTER sse2 | |
ADD_OBMC 32, sse2 | |
ADD_OBMC 16, sse2 | |
INIT_XMM sse4 | |
; void dequant_subband_32(uint8_t *src, uint8_t *dst, ptrdiff_t stride, const int qf, const int qs, int tot_v, int tot_h) | |
cglobal dequant_subband_32, 7, 7, 4, src, dst, stride, qf, qs, tot_v, tot_h | |
movd m2, qfd | |
movd m3, qsd | |
SPLATD m2 | |
SPLATD m3 | |
mov r4d, tot_hd | |
mov r3, dstq | |
.loop_v: | |
mov tot_hq, r4 | |
mov dstq, r3 | |
.loop_h: | |
movu m0, [srcq] | |
pabsd m1, m0 | |
pmulld m1, m2 | |
paddd m1, m3 | |
psrld m1, 2 | |
psignd m1, m0 | |
movu [dstq], m1 | |
add srcq, mmsize | |
add dstq, mmsize | |
sub tot_hq, 4 | |
jg .loop_h | |
lea srcq, [srcq + 4*tot_hq] | |
add r3, strideq | |
dec tot_vd | |
jg .loop_v | |
RET | |
INIT_XMM sse4 | |
; void put_signed_rect_clamped_10(uint8_t *dst, int dst_stride, const uint8_t *src, int src_stride, int width, int height) | |
%if ARCH_X86_64 | |
cglobal put_signed_rect_clamped_10, 6, 8, 5, dst, dst_stride, src, src_stride, w, h, t1, t2 | |
%else | |
cglobal put_signed_rect_clamped_10, 5, 7, 5, dst, dst_stride, src, src_stride, w, t1, t2 | |
%define hd r5mp | |
%endif | |
shl wd, 2 | |
add srcq, wq | |
neg wq | |
mov t2q, dstq | |
mov t1q, wq | |
pxor m2, m2 | |
mova m3, [clip_10bit] | |
mova m4, [convert_to_unsigned_10bit] | |
.loop_h: | |
mov dstq, t2q | |
mov wq, t1q | |
.loop_w: | |
movu m0, [srcq+wq+0*mmsize] | |
movu m1, [srcq+wq+1*mmsize] | |
paddd m0, m4 | |
paddd m1, m4 | |
packusdw m0, m0, m1 | |
CLIPW m0, m2, m3 ; packusdw saturates so it's fine | |
movu [dstq], m0 | |
add dstq, 1*mmsize | |
add wq, 2*mmsize | |
jl .loop_w | |
add srcq, src_strideq | |
add t2q, dst_strideq | |
sub hd, 1 | |
jg .loop_h | |
RET | |