ref: 7d6d86b63a66950b08fe518ddfc4466b63d73975
dir: /codec/common/arm64/copy_mb_aarch64_neon.S/
/*! * \copy * Copyright (c) 2013, Cisco Systems * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * * Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in * the documentation and/or other materials provided with the * distribution. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE * COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE * POSSIBILITY OF SUCH DAMAGE. * */ #ifdef HAVE_NEON_AARCH64 .text #include "arm_arch64_common_macro.S" #ifdef __APPLE__ .macro LOAD_ALIGNED_DATA_WITH_STRIDE // { // input: $0~$3, src*, src_stride ld1 {$0.d}[0], [$4], $5 ld1 {$1.d}[0], [$4], $5 ld1 {$2.d}[0], [$4], $5 ld1 {$3.d}[0], [$4], $5 // } .endm .macro STORE_ALIGNED_DATA_WITH_STRIDE // { // input: $0~$3, dst*, dst_stride st1 {$0.d}[0], [$4], $5 st1 {$1.d}[0], [$4], $5 st1 {$2.d}[0], [$4], $5 st1 {$3.d}[0], [$4], $5 // } .endm .macro LOAD_UNALIGNED_DATA_WITH_STRIDE // { // input: $0~$3, src*, src_stride ld1 {$0.8b}, [$4], $5 ld1 {$1.8b}, [$4], $5 ld1 {$2.8b}, [$4], $5 ld1 {$3.8b}, [$4], $5 // } .endm .macro STORE_UNALIGNED_DATA_WITH_STRIDE // { // input: $0~$3, dst*, dst_stride st1 {$0.8b}, [$4], $5 st1 {$1.8b}, [$4], $5 st1 {$2.8b}, [$4], $5 st1 {$3.8b}, [$4], $5 // } .endm .macro LOAD16_ALIGNED_DATA_WITH_STRIDE // { // input: $0~$3, src*, src_stride ld1 {$0.2d}, [$4], $5 ld1 {$1.2d}, [$4], $5 ld1 {$2.2d}, [$4], $5 ld1 {$3.2d}, [$4], $5 // } .endm .macro STORE16_ALIGNED_DATA_WITH_STRIDE // { // input: $0~$3, dst*, dst_stride st1 {$0.2d}, [$4], $5 st1 {$1.2d}, [$4], $5 st1 {$2.2d}, [$4], $5 st1 {$3.2d}, [$4], $5 // } .endm .macro LOAD16_UNALIGNED_DATA_WITH_STRIDE // { // input: $0~$3, src*, src_stride ld1 {$0.16b}, [$4], $5 ld1 {$1.16b}, [$4], $5 ld1 {$2.16b}, [$4], $5 ld1 {$3.16b}, [$4], $5 // } .endm .macro STORE16_UNALIGNED_DATA_WITH_STRIDE // { // input: $0~$3, dst*, dst_stride st1 {$0.16b}, [$4], $5 st1 {$1.16b}, [$4], $5 st1 {$2.16b}, [$4], $5 st1 {$3.16b}, [$4], $5 // } .endm #else .macro LOAD_ALIGNED_DATA_WITH_STRIDE arg0, arg1, arg2, arg3, arg4, arg5 // { // input: $0~$3, src*, src_stride ld1 {\arg0\().d}[0], [\arg4], \arg5 ld1 {\arg1\().d}[0], [\arg4], \arg5 ld1 {\arg2\().d}[0], [\arg4], \arg5 ld1 {\arg3\().d}[0], [\arg4], \arg5 // } .endm .macro STORE_ALIGNED_DATA_WITH_STRIDE arg0, arg1, arg2, arg3, arg4, arg5 // { // input: $0~$3, dst*, dst_stride st1 {\arg0\().d}[0], [\arg4], \arg5 st1 {\arg1\().d}[0], [\arg4], \arg5 st1 {\arg2\().d}[0], [\arg4], \arg5 st1 {\arg3\().d}[0], [\arg4], \arg5 // } .endm .macro LOAD_UNALIGNED_DATA_WITH_STRIDE arg0, arg1, arg2, arg3, arg4, arg5 // { // input: $0~$3, src*, src_stride ld1 {\arg0\().8b}, [\arg4], \arg5 ld1 {\arg1\().8b}, [\arg4], \arg5 ld1 {\arg2\().8b}, [\arg4], \arg5 ld1 {\arg3\().8b}, [\arg4], \arg5 // } .endm .macro STORE_UNALIGNED_DATA_WITH_STRIDE arg0, arg1, arg2, arg3, arg4, arg5 // { // input: $0~$3, dst*, dst_stride st1 {\arg0\().8b}, [\arg4], \arg5 st1 {\arg1\().8b}, [\arg4], \arg5 st1 {\arg2\().8b}, [\arg4], \arg5 st1 {\arg3\().8b}, [\arg4], \arg5 // } .endm .macro LOAD16_ALIGNED_DATA_WITH_STRIDE arg0, arg1, arg2, arg3, arg4, arg5 // { // input: $0~$3, src*, src_stride ld1 {\arg0\().2d}, [\arg4], \arg5 ld1 {\arg1\().2d}, [\arg4], \arg5 ld1 {\arg2\().2d}, [\arg4], \arg5 ld1 {\arg3\().2d}, [\arg4], \arg5 // } .endm .macro STORE16_ALIGNED_DATA_WITH_STRIDE arg0, arg1, arg2, arg3, arg4, arg5 // { // input: $0~$3, dst*, dst_stride st1 {\arg0\().2d}, [\arg4], \arg5 st1 {\arg1\().2d}, [\arg4], \arg5 st1 {\arg2\().2d}, [\arg4], \arg5 st1 {\arg3\().2d}, [\arg4], \arg5 // } .endm .macro LOAD16_UNALIGNED_DATA_WITH_STRIDE arg0, arg1, arg2, arg3, arg4, arg5 // { // input: $0~$3, src*, src_stride ld1 {\arg0\().16b}, [\arg4], \arg5 ld1 {\arg1\().16b}, [\arg4], \arg5 ld1 {\arg2\().16b}, [\arg4], \arg5 ld1 {\arg3\().16b}, [\arg4], \arg5 // } .endm .macro STORE16_UNALIGNED_DATA_WITH_STRIDE arg0, arg1, arg2, arg3, arg4, arg5 // { // input: $0~$3, dst*, dst_stride st1 {\arg0\().16b}, [\arg4], \arg5 st1 {\arg1\().16b}, [\arg4], \arg5 st1 {\arg2\().16b}, [\arg4], \arg5 st1 {\arg3\().16b}, [\arg4], \arg5 // } .endm #endif WELS_ASM_AARCH64_FUNC_BEGIN WelsCopy8x8_AArch64_neon LOAD_UNALIGNED_DATA_WITH_STRIDE v0, v1, v2, v3, x2, x3 STORE_UNALIGNED_DATA_WITH_STRIDE v0, v1, v2, v3, x0, x1 LOAD_UNALIGNED_DATA_WITH_STRIDE v4, v5, v6, v7, x2, x3 STORE_UNALIGNED_DATA_WITH_STRIDE v4, v5, v6, v7, x0, x1 WELS_ASM_AARCH64_FUNC_END WELS_ASM_AARCH64_FUNC_BEGIN WelsCopy16x16_AArch64_neon LOAD16_ALIGNED_DATA_WITH_STRIDE v0, v1, v2, v3, x2, x3 STORE16_ALIGNED_DATA_WITH_STRIDE v0, v1, v2, v3, x0, x1 LOAD16_ALIGNED_DATA_WITH_STRIDE v16, v17, v18, v19, x2, x3 STORE16_ALIGNED_DATA_WITH_STRIDE v16, v17, v18, v19, x0, x1 LOAD16_ALIGNED_DATA_WITH_STRIDE v0, v1, v2, v3, x2, x3 STORE16_ALIGNED_DATA_WITH_STRIDE v0, v1, v2, v3, x0, x1 LOAD16_ALIGNED_DATA_WITH_STRIDE v16, v17, v18, v19, x2, x3 STORE16_ALIGNED_DATA_WITH_STRIDE v16, v17, v18, v19, x0, x1 WELS_ASM_AARCH64_FUNC_END WELS_ASM_AARCH64_FUNC_BEGIN WelsCopy16x16NotAligned_AArch64_neon LOAD16_UNALIGNED_DATA_WITH_STRIDE v0, v1, v2, v3, x2, x3 STORE16_UNALIGNED_DATA_WITH_STRIDE v0, v1, v2, v3, x0, x1 LOAD16_UNALIGNED_DATA_WITH_STRIDE v16, v17, v18, v19, x2, x3 STORE16_UNALIGNED_DATA_WITH_STRIDE v16, v17, v18, v19, x0, x1 LOAD16_UNALIGNED_DATA_WITH_STRIDE v0, v1, v2, v3, x2, x3 STORE16_UNALIGNED_DATA_WITH_STRIDE v0, v1, v2, v3, x0, x1 LOAD16_UNALIGNED_DATA_WITH_STRIDE v16, v17, v18, v19, x2, x3 STORE16_UNALIGNED_DATA_WITH_STRIDE v16, v17, v18, v19, x0, x1 WELS_ASM_AARCH64_FUNC_END WELS_ASM_AARCH64_FUNC_BEGIN WelsCopy16x8NotAligned_AArch64_neon LOAD16_UNALIGNED_DATA_WITH_STRIDE v0, v1, v2, v3, x2, x3 STORE16_UNALIGNED_DATA_WITH_STRIDE v0, v1, v2, v3, x0, x1 LOAD16_UNALIGNED_DATA_WITH_STRIDE v16, v17, v18, v19, x2, x3 STORE16_UNALIGNED_DATA_WITH_STRIDE v16, v17, v18, v19, x0, x1 WELS_ASM_AARCH64_FUNC_END WELS_ASM_AARCH64_FUNC_BEGIN WelsCopy8x16_AArch64_neon LOAD_UNALIGNED_DATA_WITH_STRIDE v0, v1, v2, v3, x2, x3 STORE_UNALIGNED_DATA_WITH_STRIDE v0, v1, v2, v3, x0, x1 LOAD_UNALIGNED_DATA_WITH_STRIDE v4, v5, v6, v7, x2, x3 STORE_UNALIGNED_DATA_WITH_STRIDE v4, v5, v6, v7, x0, x1 LOAD_UNALIGNED_DATA_WITH_STRIDE v0, v1, v2, v3, x2, x3 STORE_UNALIGNED_DATA_WITH_STRIDE v0, v1, v2, v3, x0, x1 LOAD_UNALIGNED_DATA_WITH_STRIDE v4, v5, v6, v7, x2, x3 STORE_UNALIGNED_DATA_WITH_STRIDE v4, v5, v6, v7, x0, x1 WELS_ASM_AARCH64_FUNC_END #endif