1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-12-23 12:43:46 +02:00

ARM asm for AV_RN*()

ARMv6 and later support unaligned loads and stores for single
word/halfword but not double/multiple.  GCC is ignorant of this and
will always use bytewise accesses for unaligned data.  Casting to an
int32_t pointer is dangerous since a load/store double or multiple
instruction might be used (this happens with some code in FFmpeg).
Implementing the AV_[RW]* macros with inline asm using only supported
instructions gives fast and safe unaligned accesses.  ARM RVCT does
the right thing with generic code.

This gives an overall speedup of up to 10%.

Originally committed as revision 18601 to svn://svn.ffmpeg.org/ffmpeg/trunk
This commit is contained in:
Måns Rullgård 2009-04-18 00:00:28 +00:00
parent a6783b8961
commit 3c55ce039d
2 changed files with 81 additions and 0 deletions

View File

@ -0,0 +1,78 @@
/*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#ifndef AVUTIL_ARM_INTREADWRITE_H
#define AVUTIL_ARM_INTREADWRITE_H
#include <stdint.h>
#include "config.h"
#if HAVE_FAST_UNALIGNED && HAVE_INLINE_ASM
#define AV_RN16 AV_RN16
static inline uint16_t AV_RN16(const void *p)
{
uint16_t v;
__asm__ ("ldrh %0, %1" : "=r"(v) : "m"(*(const uint16_t *)p));
return v;
}
#define AV_WN16 AV_WN16
static inline void AV_WN16(void *p, uint16_t v)
{
__asm__ ("strh %1, %0" : "=m"(*(uint16_t *)p) : "r"(v));
}
#define AV_RN32 AV_RN32
static inline uint32_t AV_RN32(const void *p)
{
uint32_t v;
__asm__ ("ldr %0, %1" : "=r"(v) : "m"(*(const uint32_t *)p));
return v;
}
#define AV_WN32 AV_WN32
static inline void AV_WN32(void *p, uint32_t v)
{
__asm__ ("str %1, %0" : "=m"(*(uint32_t *)p) : "r"(v));
}
#define AV_RN64 AV_RN64
static inline uint64_t AV_RN64(const void *p)
{
union { uint64_t v; uint32_t hl[2]; } v;
__asm__ ("ldr %0, %2 \n\t"
"ldr %1, %3 \n\t"
: "=r"(v.hl[0]), "=r"(v.hl[1])
: "m"(*(const uint32_t*)p), "m"(*((const uint32_t*)p+1)));
return v.v;
}
#define AV_WN64 AV_WN64
static inline void AV_WN64(void *p, uint64_t v)
{
union { uint64_t v; uint32_t hl[2]; } vv = { v };
__asm__ ("str %2, %0 \n\t"
"str %3, %1 \n\t"
: "=m"(*(uint32_t*)p), "=m"(*((uint32_t*)p+1))
: "r"(vv.hl[0]), "r"(vv.hl[1]));
}
#endif /* HAVE_INLINE_ASM */
#endif /* AVUTIL_ARM_INTREADWRITE_H */

View File

@ -29,6 +29,9 @@
* defined, even if these are implemented as inline functions.
*/
#if ARCH_ARM
# include "arm/intreadwrite.h"
#endif
/*
* Define AV_[RW]N helper macros to simplify definitions not provided