mirror of
https://gitlab.freedesktop.org/pipewire/pipewire.git
synced 2026-03-03 01:40:51 -05:00
spa: use the right AVX2 flags
Our AVX optimizations are really AVX2 so rename the files and functions and use the right HAVE_AVX2 and cpu flags to compile and select the right functions. Fixes #5072
This commit is contained in:
parent
5871f88b81
commit
8eb60b132c
15 changed files with 66 additions and 66 deletions
74
spa/plugins/audioconvert/resample-native-avx2.c
Normal file
74
spa/plugins/audioconvert/resample-native-avx2.c
Normal file
|
|
@ -0,0 +1,74 @@
|
|||
/* Spa */
|
||||
/* SPDX-FileCopyrightText: Copyright © 2019 Wim Taymans */
|
||||
/* SPDX-License-Identifier: MIT */
|
||||
|
||||
#include "resample-native-impl.h"
|
||||
|
||||
#include <assert.h>
|
||||
#include <immintrin.h>
|
||||
|
||||
static inline void inner_product_avx2(float *d, const float * SPA_RESTRICT s,
|
||||
const float * SPA_RESTRICT taps, uint32_t n_taps)
|
||||
{
|
||||
__m256 sy[2] = { _mm256_setzero_ps(), _mm256_setzero_ps() }, ty;
|
||||
__m128 sx[2], tx;
|
||||
uint32_t i = 0;
|
||||
uint32_t n_taps4 = n_taps & ~0xf;
|
||||
|
||||
for (; i < n_taps4; i += 16) {
|
||||
ty = _mm256_loadu_ps(s + i + 0);
|
||||
sy[0] = _mm256_fmadd_ps(ty, _mm256_load_ps(taps + i + 0), sy[0]);
|
||||
ty = _mm256_loadu_ps(s + i + 8);
|
||||
sy[1] = _mm256_fmadd_ps(ty, _mm256_load_ps(taps + i + 8), sy[1]);
|
||||
}
|
||||
sy[0] = _mm256_add_ps(sy[1], sy[0]);
|
||||
sx[1] = _mm256_extractf128_ps(sy[0], 1);
|
||||
sx[0] = _mm256_extractf128_ps(sy[0], 0);
|
||||
for (; i < n_taps; i += 8) {
|
||||
tx = _mm_loadu_ps(s + i + 0);
|
||||
sx[0] = _mm_fmadd_ps(tx, _mm_load_ps(taps + i + 0), sx[0]);
|
||||
tx = _mm_loadu_ps(s + i + 4);
|
||||
sx[1] = _mm_fmadd_ps(tx, _mm_load_ps(taps + i + 4), sx[1]);
|
||||
}
|
||||
sx[0] = _mm_add_ps(sx[0], sx[1]);
|
||||
sx[0] = _mm_hadd_ps(sx[0], sx[0]);
|
||||
sx[0] = _mm_hadd_ps(sx[0], sx[0]);
|
||||
_mm_store_ss(d, sx[0]);
|
||||
}
|
||||
|
||||
static inline void inner_product_ip_avx2(float *d, const float * SPA_RESTRICT s,
|
||||
const float * SPA_RESTRICT t0, const float * SPA_RESTRICT t1, float x,
|
||||
uint32_t n_taps)
|
||||
{
|
||||
__m256 sy[2] = { _mm256_setzero_ps(), _mm256_setzero_ps() }, ty;
|
||||
__m128 sx[2], tx;
|
||||
uint32_t i, n_taps4 = n_taps & ~0xf;
|
||||
|
||||
for (i = 0; i < n_taps4; i += 16) {
|
||||
ty = _mm256_loadu_ps(s + i + 0);
|
||||
sy[0] = _mm256_fmadd_ps(ty, _mm256_load_ps(t0 + i + 0), sy[0]);
|
||||
sy[1] = _mm256_fmadd_ps(ty, _mm256_load_ps(t1 + i + 0), sy[1]);
|
||||
ty = _mm256_loadu_ps(s + i + 8);
|
||||
sy[0] = _mm256_fmadd_ps(ty, _mm256_load_ps(t0 + i + 8), sy[0]);
|
||||
sy[1] = _mm256_fmadd_ps(ty, _mm256_load_ps(t1 + i + 8), sy[1]);
|
||||
}
|
||||
sx[0] = _mm_add_ps(_mm256_extractf128_ps(sy[0], 0), _mm256_extractf128_ps(sy[0], 1));
|
||||
sx[1] = _mm_add_ps(_mm256_extractf128_ps(sy[1], 0), _mm256_extractf128_ps(sy[1], 1));
|
||||
|
||||
for (; i < n_taps; i += 8) {
|
||||
tx = _mm_loadu_ps(s + i + 0);
|
||||
sx[0] = _mm_fmadd_ps(tx, _mm_load_ps(t0 + i + 0), sx[0]);
|
||||
sx[1] = _mm_fmadd_ps(tx, _mm_load_ps(t1 + i + 0), sx[1]);
|
||||
tx = _mm_loadu_ps(s + i + 4);
|
||||
sx[0] = _mm_fmadd_ps(tx, _mm_load_ps(t0 + i + 4), sx[0]);
|
||||
sx[1] = _mm_fmadd_ps(tx, _mm_load_ps(t1 + i + 4), sx[1]);
|
||||
}
|
||||
sx[1] = _mm_mul_ps(_mm_sub_ps(sx[1], sx[0]), _mm_load1_ps(&x));
|
||||
sx[0] = _mm_add_ps(sx[0], sx[1]);
|
||||
sx[0] = _mm_hadd_ps(sx[0], sx[0]);
|
||||
sx[0] = _mm_hadd_ps(sx[0], sx[0]);
|
||||
_mm_store_ss(d, sx[0]);
|
||||
}
|
||||
|
||||
MAKE_RESAMPLER_FULL(avx2);
|
||||
MAKE_RESAMPLER_INTER(avx2);
|
||||
Loading…
Add table
Add a link
Reference in a new issue