SHA256
1
0
forked from jengelh/ffmpeg-5

Accepting request 1115537 from home:manfred-h

- Add ffmpeg-x86-mathops.patch to allow building for Leap

OBS-URL: https://build.opensuse.org/request/show/1115537
OBS-URL: https://build.opensuse.org/package/show/multimedia:libs/ffmpeg-5?expand=0&rev=71
This commit is contained in:
Jan Engelhardt 2023-10-04 09:06:02 +00:00 committed by Git OBS Bridge
parent 8097484d43
commit 581cca54e9
3 changed files with 66 additions and 0 deletions

View File

@ -1,3 +1,8 @@
-------------------------------------------------------------------
Wed Oct 4 07:59:01 UTC 2023 - Manfred Hollstein <manfred.h@gmx.net>
- Add ffmpeg-x86-mathops.patch to allow building for Leap
-------------------------------------------------------------------
Thu Jun 29 12:26:41 UTC 2023 - Jan Engelhardt <jengelh@inai.de>

View File

@ -112,6 +112,7 @@ Patch13: 0003-avcodec-libsvtav1-only-set-max_buf_sz-if-both-bitrat.patch
Patch14: 0004-avcodec-libsvtav1-use-larger-of-bit-rate-and-max-rat.patch
Patch91: ffmpeg-dlopen-openh264.patch
Patch93: soname.diff
Patch94: ffmpeg-x86-mathops.patch
%if %{with amf_sdk}
BuildRequires: AMF-devel
@ -842,6 +843,7 @@ Patch9: ffmpeg-4.4-CVE-2020-22046.patch
Patch10: ffmpeg-chromium.patch
Patch91: ffmpeg-dlopen-openh264.patch
Patch93: soname.diff
Patch94: ffmpeg-x86-mathops.patch
BuildRequires: c_compiler
Requires: this-is-only-for-build-envs

59
ffmpeg-x86-mathops.patch Normal file
View File

@ -0,0 +1,59 @@
diff -rup a/libavcodec/x86/mathops.h b/libavcodec/x86/mathops.h
--- a/libavcodec/x86/mathops.h 2023-10-01 13:02:26.829463017 +0200
+++ b/libavcodec/x86/mathops.h 2023-10-01 13:05:19.219502582 +0200
@@ -35,12 +35,20 @@
static av_always_inline av_const int MULL(int a, int b, unsigned shift)
{
int rt, dummy;
+ if (__builtin_constant_p(shift))
__asm__ (
"imull %3 \n\t"
"shrdl %4, %%edx, %%eax \n\t"
:"=a"(rt), "=d"(dummy)
- :"a"(a), "rm"(b), "ci"((uint8_t)shift)
+ :"a"(a), "rm"(b), "i"(shift & 0x1F)
);
+ else
+ __asm__ (
+ "imull %3 \n\t"
+ "shrdl %4, %%edx, %%eax \n\t"
+ :"=a"(rt), "=d"(dummy)
+ :"a"(a), "rm"(b), "c"((uint8_t)shift)
+ );
return rt;
}
@@ -113,19 +121,31 @@ __asm__ volatile(\
// avoid +32 for shift optimization (gcc should do that ...)
#define NEG_SSR32 NEG_SSR32
static inline int32_t NEG_SSR32( int32_t a, int8_t s){
+ if (__builtin_constant_p(s))
__asm__ ("sarl %1, %0\n\t"
: "+r" (a)
- : "ic" ((uint8_t)(-s))
+ : "i" (-s & 0x1F)
);
+ else
+ __asm__ ("sarl %1, %0\n\t"
+ : "+r" (a)
+ : "c" ((uint8_t)(-s))
+ );
return a;
}
#define NEG_USR32 NEG_USR32
static inline uint32_t NEG_USR32(uint32_t a, int8_t s){
+ if (__builtin_constant_p(s))
__asm__ ("shrl %1, %0\n\t"
: "+r" (a)
- : "ic" ((uint8_t)(-s))
+ : "i" (-s & 0x1F)
);
+ else
+ __asm__ ("shrl %1, %0\n\t"
+ : "+r" (a)
+ : "c" ((uint8_t)(-s))
+ );
return a;
}