/* * Copyright (C) 2019 The Android Open Source Project * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in * the documentation and/or other materials provided with the * distribution. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS * OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. */ #include #include static inline bool __bionic_is_oryon(unsigned long hwcap) { if (!(hwcap & HWCAP_CPUID)) return false; // Extract the implementor and variant bits from MIDR_EL1. // https://www.kernel.org/doc/html/latest/arch/arm64/cpu-feature-registers.html#list-of-registers-with-visible-features unsigned long midr; __asm__ __volatile__("mrs %0, MIDR_EL1" : "=r"(midr)); uint16_t cpu = (midr >> 20) & 0xfff; auto make_cpu = [](unsigned implementor, unsigned variant) { return (implementor << 4) | variant; }; // Check for implementor Qualcomm's variants 0x1..0x5 (Oryon). return cpu >= make_cpu('Q', 0x1) && cpu <= make_cpu('Q', 0x5); } extern "C" { DEFINE_IFUNC_FOR(memchr) { if (arg->_hwcap2 & HWCAP2_MTE) { RETURN_FUNC(memchr_func_t, __memchr_aarch64_mte); } else { RETURN_FUNC(memchr_func_t, __memchr_aarch64); } } MEMCHR_SHIM() DEFINE_IFUNC_FOR(memcmp) { // TODO: enable the SVE version. RETURN_FUNC(memcmp_func_t, __memcmp_aarch64); } MEMCMP_SHIM() DEFINE_IFUNC_FOR(memcpy) { if (arg->_hwcap2 & HWCAP2_MOPS) { RETURN_FUNC(memcpy_func_t, __memmove_aarch64_mops); } else if (__bionic_is_oryon(arg->_hwcap)) { RETURN_FUNC(memcpy_func_t, __memcpy_aarch64_nt); } else if (arg->_hwcap & HWCAP_ASIMD) { RETURN_FUNC(memcpy_func_t, __memcpy_aarch64_simd); } else { RETURN_FUNC(memcpy_func_t, __memcpy_aarch64); } } MEMCPY_SHIM() DEFINE_IFUNC_FOR(memmove) { if (arg->_hwcap2 & HWCAP2_MOPS) { RETURN_FUNC(memmove_func_t, __memmove_aarch64_mops); } else if (__bionic_is_oryon(arg->_hwcap)) { RETURN_FUNC(memmove_func_t, __memmove_aarch64_nt); } else if (arg->_hwcap & HWCAP_ASIMD) { RETURN_FUNC(memmove_func_t, __memmove_aarch64_simd); } else { RETURN_FUNC(memmove_func_t, __memmove_aarch64); } } MEMMOVE_SHIM() DEFINE_IFUNC_FOR(memrchr) { RETURN_FUNC(memrchr_func_t, __memrchr_aarch64); } MEMRCHR_SHIM() DEFINE_IFUNC_FOR(memset) { if (arg->_hwcap2 & HWCAP2_MOPS) { RETURN_FUNC(memset_func_t, __memset_aarch64_mops); } else if (__bionic_is_oryon(arg->_hwcap)) { RETURN_FUNC(memset_func_t, __memset_aarch64_nt); } else { RETURN_FUNC(memset_func_t, __memset_aarch64); } } MEMSET_SHIM() DEFINE_IFUNC_FOR(stpcpy) { // TODO: enable the SVE version. RETURN_FUNC(stpcpy_func_t, __stpcpy_aarch64); } STPCPY_SHIM() DEFINE_IFUNC_FOR(strchr) { if (arg->_hwcap2 & HWCAP2_MTE) { RETURN_FUNC(strchr_func_t, __strchr_aarch64_mte); } else { RETURN_FUNC(strchr_func_t, __strchr_aarch64); } } STRCHR_SHIM() DEFINE_IFUNC_FOR(strchrnul) { if (arg->_hwcap2 & HWCAP2_MTE) { RETURN_FUNC(strchrnul_func_t, __strchrnul_aarch64_mte); } else { RETURN_FUNC(strchrnul_func_t, __strchrnul_aarch64); } } STRCHRNUL_SHIM() DEFINE_IFUNC_FOR(strcmp) { // TODO: enable the SVE version. RETURN_FUNC(strcmp_func_t, __strcmp_aarch64); } STRCMP_SHIM() DEFINE_IFUNC_FOR(strcpy) { // TODO: enable the SVE version. RETURN_FUNC(strcpy_func_t, __strcpy_aarch64); } STRCPY_SHIM() DEFINE_IFUNC_FOR(strlen) { if (arg->_hwcap2 & HWCAP2_MTE) { RETURN_FUNC(strlen_func_t, __strlen_aarch64_mte); } else { RETURN_FUNC(strlen_func_t, __strlen_aarch64); } } STRLEN_SHIM() DEFINE_IFUNC_FOR(strncmp) { // TODO: enable the SVE version. RETURN_FUNC(strncmp_func_t, __strncmp_aarch64); } STRNCMP_SHIM() DEFINE_IFUNC_FOR(strnlen) { // TODO: enable the SVE version. RETURN_FUNC(strnlen_func_t, __strnlen_aarch64); } STRNLEN_SHIM() DEFINE_IFUNC_FOR(strrchr) { if (arg->_hwcap2 & HWCAP2_MTE) { RETURN_FUNC(strrchr_func_t, __strrchr_aarch64_mte); } else { RETURN_FUNC(strrchr_func_t, __strrchr_aarch64); } } STRRCHR_SHIM() } // extern "C"