mirror of
https://gitlab.freedesktop.org/pipewire/pipewire.git
synced 2025-11-02 09:01:50 -05:00
Let the mixer functions accumulate the intermediate results into a larger size variable and then clamp to the final precission. This avoids distortions because of intermediate clamping. Although the access pattern of the reads are no longer sequential, the writes are sequential and we don't need to read intermediate values. Together with the avoided clamping this is probably faster overall. Add a unit test for the various cases.
169 lines
5.1 KiB
C
169 lines
5.1 KiB
C
/* Spa
|
|
*
|
|
* Copyright © 2019 Wim Taymans
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
* DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
|
|
#include <spa/utils/defs.h>
|
|
|
|
typedef struct {
|
|
#if __BYTE_ORDER == __LITTLE_ENDIAN
|
|
uint8_t v3;
|
|
uint8_t v2;
|
|
uint8_t v1;
|
|
#else
|
|
uint8_t v1;
|
|
uint8_t v2;
|
|
uint8_t v3;
|
|
#endif
|
|
} __attribute__ ((packed)) uint24_t;
|
|
|
|
typedef struct {
|
|
#if __BYTE_ORDER == __LITTLE_ENDIAN
|
|
uint8_t v3;
|
|
uint8_t v2;
|
|
int8_t v1;
|
|
#else
|
|
int8_t v1;
|
|
uint8_t v2;
|
|
uint8_t v3;
|
|
#endif
|
|
} __attribute__ ((packed)) int24_t;
|
|
|
|
static inline uint32_t u24_to_u32(uint24_t src)
|
|
{
|
|
return ((uint32_t)src.v1 << 16) | ((uint32_t)src.v2 << 8) | (uint32_t)src.v3;
|
|
}
|
|
|
|
#define U32_TO_U24(s) (uint24_t) { .v1 = (uint8_t)(((uint32_t)s) >> 16), \
|
|
.v2 = (uint8_t)(((uint32_t)s) >> 8), .v3 = (uint8_t)((uint32_t)s) }
|
|
|
|
static inline uint24_t u32_to_u24(uint32_t src)
|
|
{
|
|
return U32_TO_U24(src);
|
|
}
|
|
|
|
static inline int32_t s24_to_s32(int24_t src)
|
|
{
|
|
return ((int32_t)src.v1 << 16) | ((uint32_t)src.v2 << 8) | (uint32_t)src.v3;
|
|
}
|
|
|
|
#define S32_TO_S24(s) (int24_t) { .v1 = (int8_t)(((int32_t)s) >> 16), \
|
|
.v2 = (uint8_t)(((uint32_t)s) >> 8), .v3 = (uint8_t)((uint32_t)s) }
|
|
|
|
static inline int24_t s32_to_s24(int32_t src)
|
|
{
|
|
return S32_TO_S24(src);
|
|
}
|
|
|
|
|
|
#define S8_MIN -128
|
|
#define S8_MAX 127
|
|
#define S8_ACCUM(a,b) ((a) + (int16_t)(b))
|
|
#define S8_CLAMP(a) (int8_t)(SPA_CLAMP((a), S8_MIN, S8_MAX))
|
|
#define U8_OFFS 128
|
|
#define U8_ACCUM(a,b) ((a) + ((int16_t)(b) - U8_OFFS))
|
|
#define U8_CLAMP(a) (uint8_t)(SPA_CLAMP((a), S8_MIN, S8_MAX) + U8_OFFS)
|
|
|
|
#define S16_MIN -32768
|
|
#define S16_MAX 32767
|
|
#define S16_ACCUM(a,b) ((a) + (int32_t)(b))
|
|
#define S16_CLAMP(a) (int16_t)(SPA_CLAMP((a), S16_MIN, S16_MAX))
|
|
#define U16_OFFS 32768
|
|
#define U16_ACCUM(a,b) ((a) + ((int32_t)(b) - U16_OFFS))
|
|
#define U16_CLAMP(a) (uint16_t)(SPA_CLAMP((a), S16_MIN, S16_MAX) + U16_OFFS)
|
|
|
|
#define S24_32_MIN -8388608
|
|
#define S24_32_MAX 8388607
|
|
#define S24_32_ACCUM(a,b) ((a) + (int32_t)(b))
|
|
#define S24_32_CLAMP(a) (int32_t)(SPA_CLAMP((a), S24_32_MIN, S24_32_MAX))
|
|
#define U24_32_OFFS 8388608
|
|
#define U24_32_ACCUM(a,b) ((a) + ((int32_t)(b) - U24_32_OFFS))
|
|
#define U24_32_CLAMP(a) (uint32_t)(SPA_CLAMP((a), S24_32_MIN, S24_32_MAX) + U24_32_OFFS)
|
|
|
|
#define S24_ACCUM(a,b) S24_32_ACCUM(a, s24_to_s32(b))
|
|
#define S24_CLAMP(a) s32_to_s24(S24_32_CLAMP(a))
|
|
#define U24_ACCUM(a,b) U24_32_ACCUM(a, u24_to_u32(b))
|
|
#define U24_CLAMP(a) u32_to_u24(U24_32_CLAMP(a))
|
|
|
|
#define S32_MIN -2147483648
|
|
#define S32_MAX 2147483647
|
|
#define S32_ACCUM(a,b) ((a) + (int64_t)(b))
|
|
#define S32_CLAMP(a) (int32_t)(SPA_CLAMP((a), S32_MIN, S32_MAX))
|
|
#define U32_OFFS 2147483648
|
|
#define U32_ACCUM(a,b) ((a) + ((int64_t)(b) - U32_OFFS))
|
|
#define U32_CLAMP(a) (uint32_t)(SPA_CLAMP((a), S32_MIN, S32_MAX) + U32_OFFS)
|
|
|
|
#define F32_ACCUM(a,b) ((a) + (b))
|
|
#define F32_CLAMP(a) (a)
|
|
#define F64_ACCUM(a,b) ((a) + (b))
|
|
#define F64_CLAMP(a) (a)
|
|
|
|
struct mix_ops {
|
|
uint32_t fmt;
|
|
uint32_t n_channels;
|
|
uint32_t cpu_flags;
|
|
|
|
void (*clear) (struct mix_ops *ops, void * SPA_RESTRICT dst, uint32_t n_samples);
|
|
void (*process) (struct mix_ops *ops,
|
|
void * SPA_RESTRICT dst,
|
|
const void * SPA_RESTRICT src[], uint32_t n_src,
|
|
uint32_t n_samples);
|
|
void (*free) (struct mix_ops *ops);
|
|
|
|
const void *priv;
|
|
};
|
|
|
|
int mix_ops_init(struct mix_ops *ops);
|
|
|
|
#define mix_ops_clear(ops,...) (ops)->clear(ops, __VA_ARGS__)
|
|
#define mix_ops_process(ops,...) (ops)->process(ops, __VA_ARGS__)
|
|
#define mix_ops_free(ops) (ops)->free(ops)
|
|
|
|
#define DEFINE_FUNCTION(name,arch) \
|
|
void mix_##name##_##arch(struct mix_ops *ops, void * SPA_RESTRICT dst, \
|
|
const void * SPA_RESTRICT src[], uint32_t n_src, \
|
|
uint32_t n_samples) \
|
|
|
|
#define MIX_OPS_MAX_ALIGN 32
|
|
|
|
DEFINE_FUNCTION(s8, c);
|
|
DEFINE_FUNCTION(u8, c);
|
|
DEFINE_FUNCTION(s16, c);
|
|
DEFINE_FUNCTION(u16, c);
|
|
DEFINE_FUNCTION(s24, c);
|
|
DEFINE_FUNCTION(u24, c);
|
|
DEFINE_FUNCTION(s32, c);
|
|
DEFINE_FUNCTION(u32, c);
|
|
DEFINE_FUNCTION(s24_32, c);
|
|
DEFINE_FUNCTION(u24_32, c);
|
|
DEFINE_FUNCTION(f32, c);
|
|
DEFINE_FUNCTION(f64, c);
|
|
|
|
#if defined(HAVE_SSE)
|
|
DEFINE_FUNCTION(f32, sse);
|
|
#endif
|
|
#if defined(HAVE_SSE2)
|
|
DEFINE_FUNCTION(f64, sse2);
|
|
#endif
|
|
#if defined(HAVE_AVX)
|
|
DEFINE_FUNCTION(f32, avx);
|
|
#endif
|