/* * Copyright © 2018-2022, VideoLAN and dav1d authors * Copyright © 2018-2022, Two Orioles, LLC * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * 1. Redistributions of source code must retain the above copyright notice, this * list of conditions and the following disclaimer. * * 2. Redistributions in binary form must reproduce the above copyright notice, * this list of conditions and the following disclaimer in the documentation * and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE * DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #ifndef DAV1D_SRC_CPU_H #define DAV1D_SRC_CPU_H #include "config.h" #include "common/attributes.h" #include "dav1d/common.h" #include "dav1d/dav1d.h" #if ARCH_AARCH64 || ARCH_ARM #include "src/arm/cpu.h" #elif ARCH_PPC64LE #include "src/ppc/cpu.h" #elif ARCH_X86 #include "src/x86/cpu.h" #endif EXTERN unsigned dav1d_cpu_flags; EXTERN unsigned dav1d_cpu_flags_mask; void dav1d_init_cpu(void); DAV1D_API void dav1d_set_cpu_flags_mask(unsigned mask); int dav1d_num_logical_processors(Dav1dContext *c); static ALWAYS_INLINE unsigned dav1d_get_cpu_flags(void) { unsigned flags = dav1d_cpu_flags & dav1d_cpu_flags_mask; #if TRIM_DSP_FUNCTIONS /* Since this function is inlined, unconditionally setting a flag here will * enable dead code elimination in the calling function. */ #if ARCH_AARCH64 || ARCH_ARM #if defined(__ARM_NEON) || defined(__APPLE__) || defined(_WIN32) || ARCH_AARCH64 flags |= DAV1D_ARM_CPU_FLAG_NEON; #endif #elif ARCH_PPC64LE #if defined(__VSX__) flags |= DAV1D_PPC_CPU_FLAG_VSX; #endif #elif ARCH_X86 #if defined(__AVX512F__) && defined(__AVX512CD__) && \ defined(__AVX512BW__) && defined(__AVX512DQ__) && \ defined(__AVX512VL__) && defined(__AVX512VNNI__) && \ defined(__AVX512IFMA__) && defined(__AVX512VBMI__) && \ defined(__AVX512VBMI2__) && defined(__AVX512VPOPCNTDQ__) && \ defined(__AVX512BITALG__) && defined(__GFNI__) && \ defined(__VAES__) && defined(__VPCLMULQDQ__) flags |= DAV1D_X86_CPU_FLAG_AVX512ICL | DAV1D_X86_CPU_FLAG_AVX2 | DAV1D_X86_CPU_FLAG_SSE41 | DAV1D_X86_CPU_FLAG_SSSE3 | DAV1D_X86_CPU_FLAG_SSE2; #elif defined(__AVX2__) flags |= DAV1D_X86_CPU_FLAG_AVX2 | DAV1D_X86_CPU_FLAG_SSE41 | DAV1D_X86_CPU_FLAG_SSSE3 | DAV1D_X86_CPU_FLAG_SSE2; #elif defined(__SSE4_1__) || defined(__AVX__) flags |= DAV1D_X86_CPU_FLAG_SSE41 | DAV1D_X86_CPU_FLAG_SSSE3 | DAV1D_X86_CPU_FLAG_SSE2; #elif defined(__SSSE3__) flags |= DAV1D_X86_CPU_FLAG_SSSE3 | DAV1D_X86_CPU_FLAG_SSE2; #elif ARCH_X86_64 || defined(__SSE2__) || \ (defined(_M_IX86_FP) && _M_IX86_FP >= 2) flags |= DAV1D_X86_CPU_FLAG_SSE2; #endif #endif #endif return flags; } #endif /* DAV1D_SRC_CPU_H */