mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-11-26 19:01:44 +02:00
67b92d68c6
Prevents mixing inlined SSE instructions and AVX instructions when the compiler generates the latter. Signed-off-by: James Almer <jamrial@gmail.com>
170 lines
4.5 KiB
C
170 lines
4.5 KiB
C
/*
|
|
* Copyright (c) 2015 James Almer
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#ifndef AVUTIL_X86_INTMATH_H
|
|
#define AVUTIL_X86_INTMATH_H
|
|
|
|
#include <stdint.h>
|
|
#include <stdlib.h>
|
|
#if HAVE_FAST_CLZ
|
|
#if defined(_MSC_VER)
|
|
#include <intrin.h>
|
|
#elif defined(__INTEL_COMPILER)
|
|
#include <immintrin.h>
|
|
#endif
|
|
#endif
|
|
#include "config.h"
|
|
|
|
#if HAVE_FAST_CLZ
|
|
#if (defined(__INTEL_COMPILER) && (__INTEL_COMPILER>=1216)) || defined(_MSC_VER)
|
|
# if defined(__INTEL_COMPILER)
|
|
# define ff_log2(x) (_bit_scan_reverse((x)|1))
|
|
# else
|
|
# define ff_log2 ff_log2_x86
|
|
static av_always_inline av_const int ff_log2_x86(unsigned int v)
|
|
{
|
|
unsigned long n;
|
|
_BitScanReverse(&n, v|1);
|
|
return n;
|
|
}
|
|
# endif
|
|
# define ff_log2_16bit av_log2
|
|
|
|
#if defined(__INTEL_COMPILER) || (defined(_MSC_VER) && (_MSC_VER >= 1700) && \
|
|
(defined(__BMI__) || !defined(__clang__)))
|
|
# define ff_ctz(v) _tzcnt_u32(v)
|
|
|
|
# if ARCH_X86_64
|
|
# define ff_ctzll(v) _tzcnt_u64(v)
|
|
# else
|
|
# define ff_ctzll ff_ctzll_x86
|
|
static av_always_inline av_const int ff_ctzll_x86(long long v)
|
|
{
|
|
return ((uint32_t)v == 0) ? _tzcnt_u32((uint32_t)(v >> 32)) + 32 : _tzcnt_u32((uint32_t)v);
|
|
}
|
|
# endif
|
|
#endif /* _MSC_VER */
|
|
|
|
#endif /* __INTEL_COMPILER */
|
|
|
|
#endif /* HAVE_FAST_CLZ */
|
|
|
|
#if defined(__GNUC__)
|
|
|
|
/* Our generic version of av_popcount is faster than GCC's built-in on
|
|
* CPUs that don't support the popcnt instruction.
|
|
*/
|
|
#if defined(__POPCNT__)
|
|
#define av_popcount __builtin_popcount
|
|
#if ARCH_X86_64
|
|
#define av_popcount64 __builtin_popcountll
|
|
#endif
|
|
|
|
#endif /* __POPCNT__ */
|
|
|
|
#if defined(__BMI2__)
|
|
|
|
#if AV_GCC_VERSION_AT_LEAST(5,1)
|
|
#define av_mod_uintp2 __builtin_ia32_bzhi_si
|
|
#elif HAVE_INLINE_ASM
|
|
/* GCC releases before 5.1.0 have a broken bzhi builtin, so for those we
|
|
* implement it using inline assembly
|
|
*/
|
|
#define av_mod_uintp2 av_mod_uintp2_bmi2
|
|
static av_always_inline av_const unsigned av_mod_uintp2_bmi2(unsigned a, unsigned p)
|
|
{
|
|
if (av_builtin_constant_p(p))
|
|
return a & ((1 << p) - 1);
|
|
else {
|
|
unsigned x;
|
|
__asm__ ("bzhi %2, %1, %0 \n\t" : "=r"(x) : "rm"(a), "r"(p));
|
|
return x;
|
|
}
|
|
}
|
|
#endif /* AV_GCC_VERSION_AT_LEAST */
|
|
|
|
#endif /* __BMI2__ */
|
|
|
|
#if defined(__SSE2__) && !defined(__INTEL_COMPILER)
|
|
|
|
#define av_clipd av_clipd_sse2
|
|
static av_always_inline av_const double av_clipd_sse2(double a, double amin, double amax)
|
|
{
|
|
#if defined(ASSERT_LEVEL) && ASSERT_LEVEL >= 2
|
|
if (amin > amax) abort();
|
|
#endif
|
|
__asm__ ("maxsd %1, %0 \n\t"
|
|
"minsd %2, %0 \n\t"
|
|
: "+&x"(a) : "xm"(amin), "xm"(amax));
|
|
return a;
|
|
}
|
|
|
|
#endif /* __SSE2__ */
|
|
|
|
#if defined(__SSE__) && !defined(__INTEL_COMPILER)
|
|
|
|
#define av_clipf av_clipf_sse
|
|
static av_always_inline av_const float av_clipf_sse(float a, float amin, float amax)
|
|
{
|
|
#if defined(ASSERT_LEVEL) && ASSERT_LEVEL >= 2
|
|
if (amin > amax) abort();
|
|
#endif
|
|
__asm__ ("maxss %1, %0 \n\t"
|
|
"minss %2, %0 \n\t"
|
|
: "+&x"(a) : "xm"(amin), "xm"(amax));
|
|
return a;
|
|
}
|
|
|
|
#endif /* __SSE__ */
|
|
|
|
#if defined(__AVX__) && !defined(__INTEL_COMPILER)
|
|
|
|
#undef av_clipd
|
|
#define av_clipd av_clipd_avx
|
|
static av_always_inline av_const double av_clipd_avx(double a, double amin, double amax)
|
|
{
|
|
#if defined(ASSERT_LEVEL) && ASSERT_LEVEL >= 2
|
|
if (amin > amax) abort();
|
|
#endif
|
|
__asm__ ("vmaxsd %1, %0, %0 \n\t"
|
|
"vminsd %2, %0, %0 \n\t"
|
|
: "+&x"(a) : "xm"(amin), "xm"(amax));
|
|
return a;
|
|
}
|
|
|
|
#undef av_clipf
|
|
#define av_clipf av_clipf_avx
|
|
static av_always_inline av_const float av_clipf_avx(float a, float amin, float amax)
|
|
{
|
|
#if defined(ASSERT_LEVEL) && ASSERT_LEVEL >= 2
|
|
if (amin > amax) abort();
|
|
#endif
|
|
__asm__ ("vmaxss %1, %0, %0 \n\t"
|
|
"vminss %2, %0, %0 \n\t"
|
|
: "+&x"(a) : "xm"(amin), "xm"(amax));
|
|
return a;
|
|
}
|
|
|
|
#endif /* __AVX__ */
|
|
|
|
#endif /* __GNUC__ */
|
|
|
|
#endif /* AVUTIL_X86_INTMATH_H */
|