mirror of
https://gitlab.freedesktop.org/pulseaudio/pulseaudio.git
synced 2025-11-03 09:01:50 -05:00
This is required to make sure the capture output has sufficient energy for the AGC to do its job.
410 lines
14 KiB
C++
410 lines
14 KiB
C++
/***
|
|
This file is part of PulseAudio.
|
|
|
|
Copyright 2011 Collabora Ltd.
|
|
|
|
Contributor: Arun Raghavan <arun.raghavan@collabora.co.uk>
|
|
|
|
PulseAudio is free software; you can redistribute it and/or modify
|
|
it under the terms of the GNU Lesser General Public License as published
|
|
by the Free Software Foundation; either version 2.1 of the License,
|
|
or (at your option) any later version.
|
|
|
|
PulseAudio is distributed in the hope that it will be useful, but
|
|
WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public License
|
|
along with PulseAudio; if not, see <http://www.gnu.org/licenses/>.
|
|
***/
|
|
|
|
#ifdef HAVE_CONFIG_H
|
|
#include <config.h>
|
|
#endif
|
|
|
|
#include <pulse/cdecl.h>
|
|
|
|
PA_C_DECL_BEGIN
|
|
#include <pulsecore/core-util.h>
|
|
#include <pulsecore/modargs.h>
|
|
|
|
#include <pulse/timeval.h>
|
|
#include "echo-cancel.h"
|
|
PA_C_DECL_END
|
|
|
|
#include <webrtc/modules/audio_processing/include/audio_processing.h>
|
|
#include <webrtc/modules/interface/module_common_types.h>
|
|
#include <webrtc/system_wrappers/include/trace.h>
|
|
|
|
#define BLOCK_SIZE_US 10000
|
|
|
|
#define DEFAULT_HIGH_PASS_FILTER true
|
|
#define DEFAULT_NOISE_SUPPRESSION true
|
|
#define DEFAULT_ANALOG_GAIN_CONTROL true
|
|
#define DEFAULT_DIGITAL_GAIN_CONTROL false
|
|
#define DEFAULT_MOBILE false
|
|
#define DEFAULT_ROUTING_MODE "speakerphone"
|
|
#define DEFAULT_COMFORT_NOISE true
|
|
#define DEFAULT_DRIFT_COMPENSATION false
|
|
#define DEFAULT_EXTENDED_FILTER false
|
|
#define DEFAULT_INTELLIGIBILITY_ENHANCER false
|
|
#define DEFAULT_TRACE false
|
|
|
|
#define WEBRTC_AGC_MAX_VOLUME 255
|
|
#define WEBRTC_AGC_START_VOLUME 85
|
|
|
|
static const char* const valid_modargs[] = {
|
|
"high_pass_filter",
|
|
"noise_suppression",
|
|
"analog_gain_control",
|
|
"digital_gain_control",
|
|
"mobile",
|
|
"routing_mode",
|
|
"comfort_noise",
|
|
"drift_compensation",
|
|
"extended_filter",
|
|
"intelligibility_enhancer",
|
|
"trace",
|
|
NULL
|
|
};
|
|
|
|
static int routing_mode_from_string(const char *rmode) {
|
|
if (pa_streq(rmode, "quiet-earpiece-or-headset"))
|
|
return webrtc::EchoControlMobile::kQuietEarpieceOrHeadset;
|
|
else if (pa_streq(rmode, "earpiece"))
|
|
return webrtc::EchoControlMobile::kEarpiece;
|
|
else if (pa_streq(rmode, "loud-earpiece"))
|
|
return webrtc::EchoControlMobile::kLoudEarpiece;
|
|
else if (pa_streq(rmode, "speakerphone"))
|
|
return webrtc::EchoControlMobile::kSpeakerphone;
|
|
else if (pa_streq(rmode, "loud-speakerphone"))
|
|
return webrtc::EchoControlMobile::kLoudSpeakerphone;
|
|
else
|
|
return -1;
|
|
}
|
|
|
|
class PaWebrtcTraceCallback : public webrtc::TraceCallback {
|
|
void Print(webrtc::TraceLevel level, const char *message, int length)
|
|
{
|
|
if (level & webrtc::kTraceError || level & webrtc::kTraceCritical)
|
|
pa_log(message);
|
|
else if (level & webrtc::kTraceWarning)
|
|
pa_log_warn(message);
|
|
else if (level & webrtc::kTraceInfo)
|
|
pa_log_info(message);
|
|
else
|
|
pa_log_debug(message);
|
|
}
|
|
};
|
|
|
|
static int webrtc_volume_from_pa(pa_volume_t v)
|
|
{
|
|
return (v * WEBRTC_AGC_MAX_VOLUME) / PA_VOLUME_NORM;
|
|
}
|
|
|
|
static pa_volume_t webrtc_volume_to_pa(int v)
|
|
{
|
|
return (v * PA_VOLUME_NORM) / WEBRTC_AGC_MAX_VOLUME;
|
|
}
|
|
|
|
static void pa_webrtc_ec_fixate_spec(pa_sample_spec *rec_ss, pa_channel_map *rec_map,
|
|
pa_sample_spec *play_ss, pa_channel_map *play_map,
|
|
pa_sample_spec *out_ss, pa_channel_map *out_map)
|
|
{
|
|
rec_ss->format = PA_SAMPLE_S16NE;
|
|
play_ss->format = PA_SAMPLE_S16NE;
|
|
|
|
/* AudioProcessing expects one of the following rates */
|
|
if (rec_ss->rate >= 48000)
|
|
rec_ss->rate = 48000;
|
|
else if (rec_ss->rate >= 32000)
|
|
rec_ss->rate = 32000;
|
|
else if (rec_ss->rate >= 16000)
|
|
rec_ss->rate = 16000;
|
|
else
|
|
rec_ss->rate = 8000;
|
|
|
|
/* In int16 mode, AudioProcessing will give us the same spec we give it */
|
|
*out_ss = *rec_ss;
|
|
*out_map = *rec_map;
|
|
|
|
/* Playback stream rate needs to be the same as capture */
|
|
play_ss->rate = rec_ss->rate;
|
|
}
|
|
|
|
bool pa_webrtc_ec_init(pa_core *c, pa_echo_canceller *ec,
|
|
pa_sample_spec *rec_ss, pa_channel_map *rec_map,
|
|
pa_sample_spec *play_ss, pa_channel_map *play_map,
|
|
pa_sample_spec *out_ss, pa_channel_map *out_map,
|
|
uint32_t *nframes, const char *args) {
|
|
webrtc::AudioProcessing *apm = NULL;
|
|
webrtc::ProcessingConfig pconfig;
|
|
webrtc::Config config;
|
|
bool hpf, ns, agc, dgc, mobile, cn, ext_filter, intelligibility;
|
|
int rm = -1;
|
|
pa_modargs *ma;
|
|
bool trace = false;
|
|
|
|
if (!(ma = pa_modargs_new(args, valid_modargs))) {
|
|
pa_log("Failed to parse submodule arguments.");
|
|
goto fail;
|
|
}
|
|
|
|
hpf = DEFAULT_HIGH_PASS_FILTER;
|
|
if (pa_modargs_get_value_boolean(ma, "high_pass_filter", &hpf) < 0) {
|
|
pa_log("Failed to parse high_pass_filter value");
|
|
goto fail;
|
|
}
|
|
|
|
ns = DEFAULT_NOISE_SUPPRESSION;
|
|
if (pa_modargs_get_value_boolean(ma, "noise_suppression", &ns) < 0) {
|
|
pa_log("Failed to parse noise_suppression value");
|
|
goto fail;
|
|
}
|
|
|
|
agc = DEFAULT_ANALOG_GAIN_CONTROL;
|
|
if (pa_modargs_get_value_boolean(ma, "analog_gain_control", &agc) < 0) {
|
|
pa_log("Failed to parse analog_gain_control value");
|
|
goto fail;
|
|
}
|
|
|
|
dgc = agc ? false : DEFAULT_DIGITAL_GAIN_CONTROL;
|
|
if (pa_modargs_get_value_boolean(ma, "digital_gain_control", &dgc) < 0) {
|
|
pa_log("Failed to parse digital_gain_control value");
|
|
goto fail;
|
|
}
|
|
|
|
if (agc && dgc) {
|
|
pa_log("You must pick only one between analog and digital gain control");
|
|
goto fail;
|
|
}
|
|
|
|
mobile = DEFAULT_MOBILE;
|
|
if (pa_modargs_get_value_boolean(ma, "mobile", &mobile) < 0) {
|
|
pa_log("Failed to parse mobile value");
|
|
goto fail;
|
|
}
|
|
|
|
ec->params.drift_compensation = DEFAULT_DRIFT_COMPENSATION;
|
|
if (pa_modargs_get_value_boolean(ma, "drift_compensation", &ec->params.drift_compensation) < 0) {
|
|
pa_log("Failed to parse drift_compensation value");
|
|
goto fail;
|
|
}
|
|
|
|
if (mobile) {
|
|
if (ec->params.drift_compensation) {
|
|
pa_log("Can't use drift_compensation in mobile mode");
|
|
goto fail;
|
|
}
|
|
|
|
if ((rm = routing_mode_from_string(pa_modargs_get_value(ma, "routing_mode", DEFAULT_ROUTING_MODE))) < 0) {
|
|
pa_log("Failed to parse routing_mode value");
|
|
goto fail;
|
|
}
|
|
|
|
cn = DEFAULT_COMFORT_NOISE;
|
|
if (pa_modargs_get_value_boolean(ma, "comfort_noise", &cn) < 0) {
|
|
pa_log("Failed to parse cn value");
|
|
goto fail;
|
|
}
|
|
} else {
|
|
if (pa_modargs_get_value(ma, "comfort_noise", NULL) || pa_modargs_get_value(ma, "routing_mode", NULL)) {
|
|
pa_log("The routing_mode and comfort_noise options are only valid with mobile=true");
|
|
goto fail;
|
|
}
|
|
}
|
|
|
|
ext_filter = DEFAULT_EXTENDED_FILTER;
|
|
if (pa_modargs_get_value_boolean(ma, "extended_filter", &ext_filter) < 0) {
|
|
pa_log("Failed to parse extended_filter value");
|
|
goto fail;
|
|
}
|
|
|
|
intelligibility = DEFAULT_INTELLIGIBILITY_ENHANCER;
|
|
if (pa_modargs_get_value_boolean(ma, "intelligibility_enhancer", &intelligibility) < 0) {
|
|
pa_log("Failed to parse intelligibility_enhancer value");
|
|
goto fail;
|
|
}
|
|
|
|
if (ext_filter)
|
|
config.Set<webrtc::ExtendedFilter>(new webrtc::ExtendedFilter(true));
|
|
if (intelligibility)
|
|
pa_log_warn("The intelligibility enhancer is not currently supported");
|
|
|
|
trace = DEFAULT_TRACE;
|
|
if (pa_modargs_get_value_boolean(ma, "trace", &trace) < 0) {
|
|
pa_log("Failed to parse trace value");
|
|
goto fail;
|
|
}
|
|
|
|
if (trace) {
|
|
webrtc::Trace::CreateTrace();
|
|
webrtc::Trace::set_level_filter(webrtc::kTraceAll);
|
|
ec->params.priv.webrtc.trace_callback = new PaWebrtcTraceCallback();
|
|
webrtc::Trace::SetTraceCallback((PaWebrtcTraceCallback *) ec->params.priv.webrtc.trace_callback);
|
|
}
|
|
|
|
pa_webrtc_ec_fixate_spec(rec_ss, rec_map, play_ss, play_map, out_ss, out_map);
|
|
|
|
apm = webrtc::AudioProcessing::Create(config);
|
|
|
|
pconfig = {
|
|
webrtc::StreamConfig(rec_ss->rate, rec_ss->channels, false), /* input stream */
|
|
webrtc::StreamConfig(out_ss->rate, out_ss->channels, false), /* output stream */
|
|
webrtc::StreamConfig(play_ss->rate, play_ss->channels, false), /* reverse input stream */
|
|
webrtc::StreamConfig(play_ss->rate, play_ss->channels, false), /* reverse output stream */
|
|
};
|
|
apm->Initialize(pconfig);
|
|
|
|
if (hpf)
|
|
apm->high_pass_filter()->Enable(true);
|
|
|
|
if (!mobile) {
|
|
apm->echo_cancellation()->enable_drift_compensation(ec->params.drift_compensation);
|
|
apm->echo_cancellation()->Enable(true);
|
|
} else {
|
|
apm->echo_control_mobile()->set_routing_mode(static_cast<webrtc::EchoControlMobile::RoutingMode>(rm));
|
|
apm->echo_control_mobile()->enable_comfort_noise(cn);
|
|
apm->echo_control_mobile()->Enable(true);
|
|
}
|
|
|
|
if (ns) {
|
|
apm->noise_suppression()->set_level(webrtc::NoiseSuppression::kHigh);
|
|
apm->noise_suppression()->Enable(true);
|
|
}
|
|
|
|
if (agc || dgc) {
|
|
if (mobile && rm <= webrtc::EchoControlMobile::kEarpiece) {
|
|
/* Maybe this should be a knob, but we've got a lot of knobs already */
|
|
apm->gain_control()->set_mode(webrtc::GainControl::kFixedDigital);
|
|
ec->params.priv.webrtc.agc = false;
|
|
} else if (dgc) {
|
|
apm->gain_control()->set_mode(webrtc::GainControl::kAdaptiveDigital);
|
|
ec->params.priv.webrtc.agc = false;
|
|
} else {
|
|
apm->gain_control()->set_mode(webrtc::GainControl::kAdaptiveAnalog);
|
|
if (apm->gain_control()->set_analog_level_limits(0, WEBRTC_AGC_MAX_VOLUME) != apm->kNoError) {
|
|
pa_log("Failed to initialise AGC");
|
|
goto fail;
|
|
}
|
|
ec->params.priv.webrtc.agc = true;
|
|
}
|
|
|
|
apm->gain_control()->Enable(true);
|
|
}
|
|
|
|
apm->voice_detection()->Enable(true);
|
|
|
|
ec->params.priv.webrtc.apm = apm;
|
|
ec->params.priv.webrtc.sample_spec = *out_ss;
|
|
ec->params.priv.webrtc.blocksize = (uint64_t)pa_bytes_per_second(out_ss) * BLOCK_SIZE_US / PA_USEC_PER_SEC;
|
|
*nframes = ec->params.priv.webrtc.blocksize / pa_frame_size(out_ss);
|
|
ec->params.priv.webrtc.first = true;
|
|
|
|
pa_modargs_free(ma);
|
|
return true;
|
|
|
|
fail:
|
|
if (ma)
|
|
pa_modargs_free(ma);
|
|
if (ec->params.priv.webrtc.trace_callback) {
|
|
webrtc::Trace::ReturnTrace();
|
|
delete ((PaWebrtcTraceCallback *) ec->params.priv.webrtc.trace_callback);
|
|
} if (apm)
|
|
delete apm;
|
|
|
|
return false;
|
|
}
|
|
|
|
void pa_webrtc_ec_play(pa_echo_canceller *ec, const uint8_t *play) {
|
|
webrtc::AudioProcessing *apm = (webrtc::AudioProcessing*)ec->params.priv.webrtc.apm;
|
|
webrtc::AudioFrame play_frame;
|
|
const pa_sample_spec *ss = &ec->params.priv.webrtc.sample_spec;
|
|
|
|
play_frame.num_channels_ = ss->channels;
|
|
play_frame.sample_rate_hz_ = ss->rate;
|
|
play_frame.interleaved_ = true;
|
|
play_frame.samples_per_channel_ = ec->params.priv.webrtc.blocksize / pa_frame_size(ss);
|
|
|
|
pa_assert(play_frame.samples_per_channel_ <= webrtc::AudioFrame::kMaxDataSizeSamples);
|
|
memcpy(play_frame.data_, play, ec->params.priv.webrtc.blocksize);
|
|
|
|
apm->ProcessReverseStream(&play_frame);
|
|
|
|
/* FIXME: If ProcessReverseStream() makes any changes to the audio, such as
|
|
* applying intelligibility enhancement, those changes don't have any
|
|
* effect. This function is called at the source side, but the processing
|
|
* would have to be done in the sink to be able to feed the processed audio
|
|
* to speakers. */
|
|
}
|
|
|
|
void pa_webrtc_ec_record(pa_echo_canceller *ec, const uint8_t *rec, uint8_t *out) {
|
|
webrtc::AudioProcessing *apm = (webrtc::AudioProcessing*)ec->params.priv.webrtc.apm;
|
|
webrtc::AudioFrame out_frame;
|
|
const pa_sample_spec *ss = &ec->params.priv.webrtc.sample_spec;
|
|
pa_cvolume v;
|
|
int old_volume, new_volume;
|
|
|
|
out_frame.num_channels_ = ss->channels;
|
|
out_frame.sample_rate_hz_ = ss->rate;
|
|
out_frame.interleaved_ = true;
|
|
out_frame.samples_per_channel_ = ec->params.priv.webrtc.blocksize / pa_frame_size(ss);
|
|
|
|
pa_assert(out_frame.samples_per_channel_ <= webrtc::AudioFrame::kMaxDataSizeSamples);
|
|
memcpy(out_frame.data_, rec, ec->params.priv.webrtc.blocksize);
|
|
|
|
if (ec->params.priv.webrtc.agc) {
|
|
pa_cvolume_init(&v);
|
|
pa_echo_canceller_get_capture_volume(ec, &v);
|
|
old_volume = webrtc_volume_from_pa(pa_cvolume_avg(&v));
|
|
apm->gain_control()->set_stream_analog_level(old_volume);
|
|
}
|
|
|
|
apm->set_stream_delay_ms(0);
|
|
apm->ProcessStream(&out_frame);
|
|
|
|
if (ec->params.priv.webrtc.agc) {
|
|
if (PA_UNLIKELY(ec->params.priv.webrtc.first)) {
|
|
/* We start at a sane default volume (taken from the Chromium
|
|
* condition on the experimental AGC in audio_processing.h). This is
|
|
* needed to make sure that there's enough energy in the capture
|
|
* signal for the AGC to work */
|
|
ec->params.priv.webrtc.first = false;
|
|
new_volume = WEBRTC_AGC_START_VOLUME;
|
|
} else {
|
|
new_volume = apm->gain_control()->stream_analog_level();
|
|
}
|
|
|
|
if (old_volume != new_volume) {
|
|
pa_cvolume_set(&v, ss->channels, webrtc_volume_to_pa(new_volume));
|
|
pa_echo_canceller_set_capture_volume(ec, &v);
|
|
}
|
|
}
|
|
|
|
memcpy(out, out_frame.data_, ec->params.priv.webrtc.blocksize);
|
|
}
|
|
|
|
void pa_webrtc_ec_set_drift(pa_echo_canceller *ec, float drift) {
|
|
webrtc::AudioProcessing *apm = (webrtc::AudioProcessing*)ec->params.priv.webrtc.apm;
|
|
const pa_sample_spec *ss = &ec->params.priv.webrtc.sample_spec;
|
|
|
|
apm->echo_cancellation()->set_stream_drift_samples(drift * ec->params.priv.webrtc.blocksize / pa_frame_size(ss));
|
|
}
|
|
|
|
void pa_webrtc_ec_run(pa_echo_canceller *ec, const uint8_t *rec, const uint8_t *play, uint8_t *out) {
|
|
pa_webrtc_ec_play(ec, play);
|
|
pa_webrtc_ec_record(ec, rec, out);
|
|
}
|
|
|
|
void pa_webrtc_ec_done(pa_echo_canceller *ec) {
|
|
if (ec->params.priv.webrtc.trace_callback) {
|
|
webrtc::Trace::ReturnTrace();
|
|
delete ((PaWebrtcTraceCallback *) ec->params.priv.webrtc.trace_callback);
|
|
}
|
|
|
|
if (ec->params.priv.webrtc.apm) {
|
|
delete (webrtc::AudioProcessing*)ec->params.priv.webrtc.apm;
|
|
ec->params.priv.webrtc.apm = NULL;
|
|
}
|
|
}
|