pipewire/src/modules/module-rtp/audio.c
Carlos Rafael Giani 955c9ae837 module-rtp: Get the current stream time in a reusable manner
That way, redundant pw_stream_get_nsec() and clock_gettime()
calls can be avoided.
2025-10-27 22:40:22 +01:00

824 lines
29 KiB
C

/* PipeWire */
/* SPDX-FileCopyrightText: Copyright © 2022 Wim Taymans <wim.taymans@gmail.com> */
/* SPDX-License-Identifier: MIT */
static inline void
set_iovec(struct spa_ringbuffer *rbuf, void *buffer, uint32_t size,
uint32_t offset, struct iovec *iov, uint32_t len)
{
iov[0].iov_len = SPA_MIN(len, size - offset);
iov[0].iov_base = SPA_PTROFF(buffer, offset, void);
iov[1].iov_len = len - iov[0].iov_len;
iov[1].iov_base = buffer;
}
static void ringbuffer_clear(struct spa_ringbuffer *rbuf SPA_UNUSED,
void *buffer, uint32_t size,
uint32_t offset, uint32_t len)
{
struct iovec iov[2];
set_iovec(rbuf, buffer, size, offset, iov, len);
memset(iov[0].iov_base, 0, iov[0].iov_len);
memset(iov[1].iov_base, 0, iov[1].iov_len);
}
static void rtp_audio_process_playback(void *data)
{
struct impl *impl = data;
struct pw_buffer *buf;
struct spa_data *d;
struct pw_time pwt;
uint32_t wanted, timestamp, target_buffer, stride, maxsize;
uint32_t device_delay;
int32_t avail, flags = 0;
if ((buf = pw_stream_dequeue_buffer(impl->stream)) == NULL) {
pw_log_info("Out of stream buffers: %m");
return;
}
d = buf->buffer->datas;
stride = impl->stride;
maxsize = d[0].maxsize / stride;
wanted = buf->requested ? SPA_MIN(buf->requested, maxsize) : maxsize;
pw_stream_get_time_n(impl->stream, &pwt, sizeof(pwt));
/* Negative delay is used rarely, mostly for the combine stream.
* There, the delay is used as an offset value between streams.
* Here, negative delay values make no sense. It is safe to clamp
* delay values to 0 (see docs), so do that here. */
device_delay = SPA_MAX(pwt.delay, 0LL);
/* IMPORTANT: In the explanations below, sometimes, "reading/writing from/to the
* ring buffer at a position X" is mentioned. To be exact, that buffer is actually
* impl->buffer. And since X can be a timestamp whose value is far higher than the
* buffer size (and the fact that impl->buffer is a _ring_ buffer), reads and writes
* actually first do a modulo operation to the position to implement a ring buffer
* index wrap-around. (Wrap-around when reading / writing the data bytes is
* handled by the spa_ringbuffer code; this is about the wrap around of the
* read or write index itself.) */
if (impl->direct_timestamp) {
/* In direct timestamp mode, the focus lies on synchronized playback, not
* on a constant latency. The ring buffer fill level is not of interest
* here. The code in rtp_audio_receive() writes to the ring buffer at
* position (RTP timestamp + target_buffer), just like in the constant
* latency mode. Crucially however, in direct timestamp mode, it is assumed
* that the RTP timestamps are based on the same synchronized clock that
* runs the graph driver here, so the clock position is using the same
* time base as these timestamps.
*
* If the transport delay from the sender to this receiver were zero, then
* the data with the given RTP timestamp could in theory be played right
* away, since that timestamp would equal the clock position (or, in other
* words, it would be the present time). Since the transport takes some
* time, writing the data at the position (RTP timestamp + target_buffer)
* shifts the timestamp into the future sufficiently enough that no data
* is lost. (target_buffer corresponds to the `sess.latency.msec` RTP
* source module option, and that option has to be chosen by the user
* to be of a sensible size - high enough to at least match the maximum
* transport delay, but not too high to not risk too much latency
* Also, `sess.latency.msec` must be the same value across all RTP
* source nodes that shall play in sync.)
*
* When the code here reads from the position defined by the current
* clock position, it is then guaranteed that the data is accessed in
* sync with other RTP source nodes which also run in the direct
* timestamp mode, since all of them shift the timestamp by the same
* `sess.latency.msec` into the future.
*
* "Fill level" makes no sense in this mode, since a constant latency
* is not important in this mode, so no DLL is needed. Also, matching
* the pace of the synchronized clock is done by having the graph
* driver be synchronized to that clock, which will in turn cause
* any output sinks to adjust their DLLs (or similar control loop
* mechanisms) to match the pace of their data consumption with the
* pace of the driver. */
if (impl->io_position) {
/* Shift clock position by stream delay to compensate
* for processing and output delay. */
timestamp = impl->io_position->clock.position + device_delay;
spa_ringbuffer_read_update(&impl->ring, timestamp);
} else {
/* In the unlikely case that no spa_io_position pointer
* was passed yet by PipeWire to this node, resort to a
* default behavior: just use the current read index.
* This most likely is not in sync with other nodes,
* but _something_ is needed as read index until the
* spa_io_position is available. */
spa_ringbuffer_get_read_index(&impl->ring, &timestamp);
}
spa_ringbuffer_read_data(&impl->ring,
impl->buffer,
impl->actual_max_buffer_size,
(timestamp * stride) % impl->actual_max_buffer_size,
d[0].data, wanted * stride);
/* Clear the bytes that were just retrieved. Since the fill level
* is not tracked in this buffer mode, it is possible that as soon
* as actual playback ends, the RTP source node re-reads old data.
* Make sure it reads silence when no actual new data is present
* and the RTP source node still runs. Do this by filling the
* region of the retrieved data with null bytes. */
ringbuffer_clear(&impl->ring,
impl->buffer,
impl->actual_max_buffer_size,
(timestamp * stride) % impl->actual_max_buffer_size,
wanted * stride);
if (!impl->io_position) {
/* In the unlikely case that no spa_io_position pointer
* was passed yet by PipeWire to this node, monotonically
* increment the read index like this to not consume from
* the same position in the ring buffer over and over again. */
timestamp += wanted;
spa_ringbuffer_read_update(&impl->ring, timestamp);
}
} else {
/* In the constant delay mode, it is assumed that the ring buffer fill
* level matches impl->target_buffer. If not, check for over- and
* underruns. Adjust the DLL as needed. If the over/underruns are too
* severe, resynchronize. */
avail = spa_ringbuffer_get_read_index(&impl->ring, &timestamp);
/* Reduce target buffer by the delay amount to start playback sooner.
* This compensates for the delay to the device. */
if (SPA_UNLIKELY(impl->target_buffer < device_delay)) {
pw_log_error("Delay to device (%" PRIu32 ") is higher than "
"the target buffer size (%" PRIu32 ")", device_delay,
impl->target_buffer);
target_buffer = 0;
} else {
target_buffer = impl->target_buffer - device_delay;
}
if (avail < (int32_t)wanted) {
enum spa_log_level level;
memset(d[0].data, 0, wanted * stride);
flags |= SPA_CHUNK_FLAG_EMPTY;
if (impl->have_sync) {
impl->have_sync = false;
level = SPA_LOG_LEVEL_INFO;
} else {
level = SPA_LOG_LEVEL_DEBUG;
}
pw_log(level, "receiver read underrun %d/%u < %u",
avail, target_buffer, wanted);
} else {
double error, corr;
if (impl->first) {
if ((uint32_t)avail > target_buffer) {
uint32_t skip = avail - target_buffer;
pw_log_debug("first: avail:%d skip:%u target:%u",
avail, skip, target_buffer);
timestamp += skip;
avail = target_buffer;
}
impl->first = false;
} else if (avail > (int32_t)SPA_MIN(target_buffer * 8, BUFFER_SIZE / stride)) {
pw_log_warn("receiver read overrun %u > %u", avail, target_buffer * 8);
timestamp += avail - target_buffer;
avail = target_buffer;
}
/* when the speed of the sender clock and our clock are
* not in sync, try to adjust our playback rate to keep
* the requested target_buffer bytes in the ringbuffer */
double in_flight = 0;
struct spa_io_position *pos = impl->io_position;
if (SPA_LIKELY(pos && impl->last_recv_timestamp)) {
/* Account for samples that might be in flight but not yet received, and possibly
* samples that were received _after_ the process() tick and therefore should not
* yet be accounted for */
int64_t in_flight_ns = pos->clock.nsec - impl->last_recv_timestamp;
/* Use the best relative rate we know */
double relative_rate = impl->io_rate_match ? impl->io_rate_match->rate : pos->clock.rate_diff;
in_flight = (double)(in_flight_ns * impl->rate) * relative_rate / SPA_NSEC_PER_SEC;
}
error = (double)target_buffer - (double)avail - in_flight;
error = SPA_CLAMPD(error, -impl->max_error, impl->max_error);
corr = spa_dll_update(&impl->dll, error);
pw_log_trace("avail:%u target:%u error:%f corr:%f", avail,
target_buffer, error, corr);
pw_stream_set_rate(impl->stream, 1.0 / corr);
spa_ringbuffer_read_data(&impl->ring,
impl->buffer,
impl->actual_max_buffer_size,
(timestamp * stride) % impl->actual_max_buffer_size,
d[0].data, wanted * stride);
timestamp += wanted;
spa_ringbuffer_read_update(&impl->ring, timestamp);
}
}
d[0].chunk->offset = 0;
d[0].chunk->size = wanted * stride;
d[0].chunk->stride = stride;
d[0].chunk->flags = flags;
buf->size = wanted;
pw_stream_queue_buffer(impl->stream, buf);
}
static int rtp_audio_receive(struct impl *impl, uint8_t *buffer, ssize_t len,
uint64_t current_time)
{
struct rtp_header *hdr;
ssize_t hlen, plen;
uint16_t seq;
uint32_t timestamp, samples, write, expected_write;
uint32_t stride = impl->stride;
int32_t filled;
if (len < 12)
goto short_packet;
hdr = (struct rtp_header*)buffer;
if (hdr->v != 2)
goto invalid_version;
hlen = 12 + hdr->cc * 4;
if (hlen > len)
goto invalid_len;
if (impl->have_ssrc && impl->ssrc != hdr->ssrc)
goto unexpected_ssrc;
impl->ssrc = hdr->ssrc;
impl->have_ssrc = !impl->ignore_ssrc;
seq = ntohs(hdr->sequence_number);
if (impl->have_seq && impl->seq != seq) {
pw_log_info("unexpected seq (%d != %d) SSRC:%u",
seq, impl->seq, hdr->ssrc);
/* No need to resynchronize here. If packets arrive out of
* order, then they are still written in order into the ring
* buffer, since they are written according to where the
* RTP timestamp points to. */
}
impl->seq = seq + 1;
impl->have_seq = true;
timestamp = ntohl(hdr->timestamp) - impl->ts_offset;
impl->receiving = true;
impl->last_recv_timestamp = current_time;
plen = len - hlen;
samples = plen / stride;
filled = spa_ringbuffer_get_write_index(&impl->ring, &expected_write);
/* we always write to timestamp + delay */
write = timestamp + impl->target_buffer;
if (!impl->have_sync) {
pw_log_info("sync to timestamp:%u seq:%u ts_offset:%u SSRC:%u target:%u direct:%u",
timestamp, seq, impl->ts_offset, impl->ssrc,
impl->target_buffer, impl->direct_timestamp);
/* we read from timestamp, keeping target_buffer of data
* in the ringbuffer. */
impl->ring.readindex = timestamp;
impl->ring.writeindex = write;
filled = impl->target_buffer;
spa_dll_init(&impl->dll);
spa_dll_set_bw(&impl->dll, SPA_DLL_BW_MIN, 128, impl->rate);
memset(impl->buffer, 0, BUFFER_SIZE);
impl->have_sync = true;
} else if (expected_write != write) {
pw_log_debug("unexpected write (%u != %u)",
write, expected_write);
}
/* Write overrun only makes sense in constant delay mode. See the
* RTP source module documentation and the rtp_audio_process_playback()
* code for an explanation why. */
if (!impl->direct_timestamp && (filled + samples > BUFFER_SIZE / stride)) {
pw_log_debug("receiver write overrun %u + %u > %u", filled, samples,
BUFFER_SIZE / stride);
impl->have_sync = false;
} else {
pw_log_trace("got samples:%u", samples);
spa_ringbuffer_write_data(&impl->ring,
impl->buffer,
impl->actual_max_buffer_size,
(write * stride) % impl->actual_max_buffer_size,
&buffer[hlen], (samples * stride));
/* Only update the write index if data was actually _appended_.
* If packets arrived out of order, then it may be that parts
* of the ring buffer further ahead were written to first, and
* now, unwritten parts preceding those other parts were now
* written to. For example, if previously, 10 samples were
* written to index 100, even though 10 samples were expected
* to be written at index 90, then there is a "hole" at index
* 90. If now, the packet that contains data for index 90
* arrived, then this data will be _inserted_ at index 90,
* and not _appended_. In this example, `expected_write` would
* be 100 (since `expected_write` is the current write index),
* `write` would be 90, `samples` would be 10. In this case,
* the inequality below does not hold, so data is being
* _inserted_. By contrast, during normal operation, `write`
* and `expected_write` are equal, so the inequality below
* _does_ hold, meaning that data is being appended.
*
* (Note that this write index update is only important if
* the constant delay mode is active, or if no spa_io_position
* was not provided yet. See the rtp_audio_process_playback()
* code for more about this.) */
if (expected_write < (write + samples)) {
write += samples;
spa_ringbuffer_write_update(&impl->ring, write);
}
}
return 0;
short_packet:
pw_log_warn("short packet received");
return -EINVAL;
invalid_version:
pw_log_warn("invalid RTP version");
spa_debug_log_mem(pw_log_get(), SPA_LOG_LEVEL_INFO, 0, buffer, len);
return -EPROTO;
invalid_len:
pw_log_warn("invalid RTP length");
return -EINVAL;
unexpected_ssrc:
if (!impl->fixed_ssrc) {
/* We didn't have a configured SSRC, and there's more than one SSRC on
* this address/port pair */
pw_log_warn("unexpected SSRC (expected %u != %u)", impl->ssrc,
hdr->ssrc);
}
return -EINVAL;
}
static void set_timer(struct impl *impl, uint64_t time, uint64_t itime)
{
struct itimerspec ts;
ts.it_value.tv_sec = time / SPA_NSEC_PER_SEC;
ts.it_value.tv_nsec = time % SPA_NSEC_PER_SEC;
ts.it_interval.tv_sec = itime / SPA_NSEC_PER_SEC;
ts.it_interval.tv_nsec = itime % SPA_NSEC_PER_SEC;
spa_system_timerfd_settime(impl->data_loop->system,
impl->timer->fd, SPA_FD_TIMER_ABSTIME, &ts, NULL);
set_timer_running(impl, time != 0 && itime != 0);
}
static void rtp_audio_flush_packets(struct impl *impl, uint32_t num_packets, uint64_t set_timestamp)
{
int32_t avail, tosend;
uint32_t stride, timestamp;
struct iovec iov[3];
struct rtp_header header;
bool insufficient_data;
avail = spa_ringbuffer_get_read_index(&impl->ring, &timestamp);
tosend = impl->psamples;
insufficient_data = (avail < tosend);
if (insufficient_data) {
/* There is insufficient data for even a single full packet.
* Handle this depending on the current state. */
if (get_internal_stream_state(impl) == RTP_STREAM_INTERNAL_STATE_STARTED) {
/* If the stream is started, just try again later,
* when more data comes in. Enough data for covering
* the psamples amount might be available by then. */
goto done;
} else {
/* There is not enough data for a full packet, but the
* stream is no longer in the started state, so the
* remaining data needs to be flushed out now. */
tosend = avail;
num_packets = 1;
}
} else {
/* There is sufficient data for one or more full packets. */
num_packets = SPA_MIN(num_packets, (uint32_t)(avail / tosend));
}
stride = impl->stride;
spa_zero(header);
header.v = 2;
header.pt = impl->payload;
header.ssrc = htonl(impl->ssrc);
iov[0].iov_base = &header;
iov[0].iov_len = sizeof(header);
while (num_packets > 0) {
if (impl->marker_on_first && impl->first)
header.m = 1;
else
header.m = 0;
header.sequence_number = htons(impl->seq);
header.timestamp = htonl(impl->ts_offset + (set_timestamp ? set_timestamp : timestamp));
set_iovec(&impl->ring,
impl->buffer, impl->actual_max_buffer_size,
(timestamp * stride) % impl->actual_max_buffer_size,
&iov[1], tosend * stride);
pw_log_trace("sending %d packet:%d ts_offset:%d timestamp:%d",
tosend, num_packets, impl->ts_offset, timestamp);
rtp_stream_emit_send_packet(impl, iov, 3);
impl->seq++;
impl->first = false;
timestamp += tosend;
avail -= tosend;
num_packets--;
}
spa_ringbuffer_read_update(&impl->ring, timestamp);
done:
if (is_timer_running(impl)) {
if (get_internal_stream_state(impl) != RTP_STREAM_INTERNAL_STATE_STOPPING) {
/* If the stream isn't being stopped, and instead is running,
* keep the timer running if there was sufficient data to
* produce at least one packet. That's because by the time
* the next timer expiration happens, there might be enough
* data available for even more packets. However, if there
* wasn't sufficient data for even one packet, stop the
* timer, since it is likely then that input has ceased
* (at least for now). */
if (insufficient_data) {
set_timer(impl, 0, 0);
}
} else if (avail <= 0) {
/* All packets were sent, and the stream is in the stopping
* state. This means that stream_stop() was called while this
* timer was still sending out remaining packets, and thus,
* stream_stop() could not immediately change the stream to the
* stopping state. Now that all packets have gone out, finish
* the stopping state change. */
set_timer(impl, 0, 0);
pw_loop_invoke(impl->main_loop, do_finish_stopping_state, SPA_ID_INVALID, NULL, 0, false, impl);
}
}
}
static void rtp_audio_stop_timer(struct impl *impl)
{
set_timer(impl, 0, 0);
}
static void rtp_audio_flush_timeout(struct impl *impl, uint64_t expirations)
{
if (expirations > 1)
pw_log_warn("missing timeout %"PRIu64, expirations);
rtp_audio_flush_packets(impl, expirations, 0);
}
static void rtp_audio_process_capture(void *data)
{
struct impl *impl = data;
struct pw_buffer *buf;
struct spa_data *d;
uint32_t offs, size, actual_timestamp, expected_timestamp, stride;
int32_t filled, wanted;
uint32_t pending, num_queued;
struct spa_io_position *pos;
uint64_t next_nsec, quantum;
if (impl->separate_sender) {
/* apply the DLL rate */
pw_stream_set_rate(impl->stream, impl->ptp_corr);
}
if ((buf = pw_stream_dequeue_buffer(impl->stream)) == NULL) {
pw_log_info("Out of stream buffers: %m");
return;
}
d = buf->buffer->datas;
offs = SPA_MIN(d[0].chunk->offset, d[0].maxsize);
size = SPA_MIN(d[0].chunk->size, d[0].maxsize - offs);
stride = impl->stride;
wanted = size / stride;
filled = spa_ringbuffer_get_write_index(&impl->ring, &expected_timestamp);
pos = impl->io_position;
if (SPA_LIKELY(pos)) {
uint32_t rate = pos->clock.rate.denom;
actual_timestamp = pos->clock.position * impl->rate / rate;
next_nsec = pos->clock.next_nsec;
quantum = (uint64_t)(pos->clock.duration * SPA_NSEC_PER_SEC / (rate * pos->clock.rate_diff));
if (impl->separate_sender) {
/* the sender process() function uses this for managing the DLL */
impl->sink_nsec = pos->clock.nsec;
impl->sink_next_nsec = pos->clock.next_nsec;
impl->sink_resamp_delay = impl->io_rate_match->delay;
impl->sink_quantum = (uint64_t)(pos->clock.duration * SPA_NSEC_PER_SEC / rate);
}
} else {
actual_timestamp = expected_timestamp;
next_nsec = 0;
quantum = 0;
}
/* First do the synchronization checks (if the sender is in sync already.) */
if (impl->have_sync) {
if (SPA_FLAG_IS_SET(pos->clock.flags, SPA_IO_CLOCK_FLAG_DISCONT)) {
pw_log_info("IO clock reports discontinuity; resynchronizing");
impl->have_sync = false;
} else if (SPA_ABS((int64_t)expected_timestamp - (int64_t)actual_timestamp) > (int64_t)(pos->clock.duration)) {
/* Normally, expected and actual timestamp should be in sync, and deviate
* only minimally at most. If a major deviation occurs, then most likely
* the driver clock has experienced an unexpected jump. Note that the
* cycle duration in samples is used, and not the value of "quantum".
* That value is given in nanoseconds, not samples. Also, the timestamps
* themselves are not affected by rate_diff. See the documentation
* "Driver architecture and workflow" for an explanation why not. */
pw_log_warn("timestamp: expected %u != actual %u", expected_timestamp, actual_timestamp);
impl->have_sync = false;
} else if (filled + wanted > (int32_t)SPA_MIN(impl->target_buffer * 8, BUFFER_SIZE / stride)) {
pw_log_warn("sender write overrun %u + %u > %u/%u", filled, wanted,
impl->target_buffer * 8, BUFFER_SIZE / stride);
impl->have_sync = false;
filled = 0;
}
}
/* Next, (re)synchronize. If the sender was in sync, but the checks above detected
* that resynchronization is needed, then this will be done immediately below. */
if (!impl->have_sync) {
pw_log_info("(re)sync to timestamp:%u seq:%u ts_offset:%u SSRC:%u",
actual_timestamp, impl->seq, impl->ts_offset, impl->ssrc);
impl->ring.readindex = impl->ring.writeindex = actual_timestamp;
memset(impl->buffer, 0, BUFFER_SIZE);
impl->have_sync = true;
expected_timestamp = actual_timestamp;
filled = 0;
if (impl->separate_sender) {
/* the sender should know that the sync state has changed, and that it should
* refill the buffer */
impl->refilling = true;
}
}
pw_log_trace("writing %u samples at %u", wanted, expected_timestamp);
spa_ringbuffer_write_data(&impl->ring,
impl->buffer,
impl->actual_max_buffer_size,
(expected_timestamp * stride) % impl->actual_max_buffer_size,
SPA_PTROFF(d[0].data, offs, void), wanted * stride);
expected_timestamp += wanted;
spa_ringbuffer_write_update(&impl->ring, expected_timestamp);
pw_stream_queue_buffer(impl->stream, buf);
if (impl->separate_sender) {
/* sending will happen in a separate process() */
return;
}
pending = filled / impl->psamples;
num_queued = (filled + wanted) / impl->psamples;
if (num_queued > 0) {
/* flush all previous packets plus new one right away */
rtp_audio_flush_packets(impl, pending + 1, 0);
num_queued -= SPA_MIN(num_queued, pending + 1);
if (num_queued > 0) {
/* schedule timer for remaining */
int64_t interval = quantum / (num_queued + 1);
uint64_t time = next_nsec - num_queued * interval;
pw_log_trace("%u %u %"PRIu64" %"PRIu64, pending, num_queued, time, interval);
set_timer(impl, time, interval);
}
}
}
static void ptp_sender_destroy(void *d)
{
struct impl *impl = d;
spa_hook_remove(&impl->ptp_sender_listener);
impl->ptp_sender = NULL;
}
static void ptp_sender_process(void *d, struct spa_io_position *position)
{
struct impl *impl = d;
uint64_t nsec, next_nsec, quantum, quantum_nsec;
uint32_t ptp_timestamp, rtp_timestamp, read_idx;
uint32_t rate;
uint32_t filled;
double error, in_flight, delay;
nsec = position->clock.nsec;
next_nsec = position->clock.next_nsec;
/* the ringbuffer indices are in sink timetamp domain */
filled = spa_ringbuffer_get_read_index(&impl->ring, &read_idx);
if (SPA_LIKELY(position)) {
rate = position->clock.rate.denom;
quantum = position->clock.duration;
quantum_nsec = (uint64_t)(quantum * SPA_NSEC_PER_SEC / rate);
/* PTP time tells us what time it is */
ptp_timestamp = position->clock.position * impl->rate / rate;
/* RTP time is based on when we sent the first packet after the last sync */
rtp_timestamp = impl->rtp_base_ts + read_idx;
} else {
pw_log_warn("No clock information, skipping");
return;
}
pw_log_trace("sink nsec:%"PRIu64", sink next_nsec:%"PRIu64", ptp nsec:%"PRIu64", ptp next_sec:%"PRIu64,
impl->sink_nsec, impl->sink_next_nsec, nsec, next_nsec);
/* If send is lagging by more than 2 or more quanta, reset */
if (!impl->refilling && impl->rtp_last_ts &&
SPA_ABS((int32_t)ptp_timestamp - (int32_t)impl->rtp_last_ts) >= (int32_t)(2 * quantum)) {
pw_log_warn("expected %u - timestamp %u = %d >= 2 * %"PRIu64" quantum", rtp_timestamp, impl->rtp_last_ts,
(int)ptp_timestamp - (int)impl->rtp_last_ts, quantum);
goto resync;
}
if (!impl->have_sync) {
pw_log_trace("Waiting for sync");
return;
}
in_flight = (double)impl->sink_quantum * impl->rate / SPA_NSEC_PER_SEC *
(double)(nsec - impl->sink_nsec) / (impl->sink_next_nsec - impl->sink_nsec);
delay = filled + in_flight + impl->sink_resamp_delay;
/* Make sure the PTP node wake up times are within the bounds of sink
* node wake up times (with a little bit of tolerance). */
if (SPA_LIKELY(nsec > impl->sink_nsec - quantum_nsec &&
nsec < impl->sink_next_nsec + quantum_nsec)) {
/* Start adjusting if we're at/past the target delay. We requested ~1/2 the buffer
* size as the sink latency, so doing so ensures that we have two sink quanta of
* data, making the chance of and underrun low even for small buffer values */
if (impl->refilling && (double)impl->target_buffer - delay <= 0) {
impl->refilling = false;
/* Store the offset for the PTP time at which we start sending */
impl->rtp_base_ts = ptp_timestamp - read_idx;
rtp_timestamp = impl->rtp_base_ts + read_idx; /* = ptp_timestamp */
pw_log_debug("start sending. sink quantum:%"PRIu64", ptp quantum:%"PRIu64"", impl->sink_quantum, quantum_nsec);
}
if (!impl->refilling) {
/*
* As per Controlling Adaptive Resampling paper[1], maintain
* W(t) - R(t) - delta = 0. We keep delta as target_buffer.
*
* [1] http://kokkinizita.linuxaudio.org/papers/adapt-resamp.pdf
*/
error = delay - impl->target_buffer;
error = SPA_CLAMPD(error, -impl->max_error, impl->max_error);
impl->ptp_corr = spa_dll_update(&impl->ptp_dll, error);
pw_log_debug("filled:%u in_flight:%g delay:%g target:%u error:%f corr:%f",
filled, in_flight, delay, impl->target_buffer, error, impl->ptp_corr);
if (filled >= impl->psamples) {
rtp_audio_flush_packets(impl, 1, rtp_timestamp);
impl->rtp_last_ts = rtp_timestamp;
}
}
} else {
pw_log_warn("PTP node wake up time out of bounds !(%"PRIu64" < %"PRIu64" < %"PRIu64")",
impl->sink_nsec, nsec, impl->sink_next_nsec);
goto resync;
}
return;
resync:
impl->have_sync = false;
impl->rtp_last_ts = 0;
return;
}
static const struct pw_filter_events ptp_sender_events = {
PW_VERSION_FILTER_EVENTS,
.destroy = ptp_sender_destroy,
.process = ptp_sender_process
};
static int setup_ptp_sender(struct impl *impl, struct pw_core *core, enum pw_direction direction, const char *driver_grp)
{
const struct spa_pod *params[4];
struct pw_properties *filter_props = NULL;
struct spa_pod_builder b;
uint32_t n_params;
uint8_t buffer[1024];
int ret;
if (direction != PW_DIRECTION_INPUT)
return 0;
if (driver_grp == NULL) {
pw_log_info("AES67 driver group not specified, no separate sender configured");
return 0;
}
pw_log_info("AES67 driver group: %s, setting up separate sender", driver_grp);
spa_dll_init(&impl->ptp_dll);
/* BW selected empirically, as it converges most quickly and holds reasonably well in testing */
spa_dll_set_bw(&impl->ptp_dll, SPA_DLL_BW_MAX, impl->psamples, impl->rate);
impl->ptp_corr = 1.0;
n_params = 0;
spa_pod_builder_init(&b, buffer, sizeof(buffer));
filter_props = pw_properties_new(NULL, NULL);
if (filter_props == NULL) {
int res = -errno;
pw_log_error( "can't create properties: %m");
return res;
}
pw_properties_set(filter_props, PW_KEY_NODE_GROUP, driver_grp);
pw_properties_setf(filter_props, PW_KEY_NODE_NAME, "%s-ptp-sender", pw_stream_get_name(impl->stream));
pw_properties_set(filter_props, PW_KEY_NODE_ALWAYS_PROCESS, "true");
/*
* sess.latency.msec defines how much data is buffered before it is
* sent out on the network. This is done by setting the node.latency
* to that value, and process function will get chunks of that size.
* It is then split up into psamples chunks and send every ptime.
*
* With this separate sender mechanism we have some latency in stream
* via node.latency, and some in ringbuffer between sink and sender.
* Ideally we want to have a total latency that still corresponds to
* sess.latency.msec. We do this by using the property setting and
* splitting some of it as stream latency and some as ringbuffer
* latency. The ringbuffer latency is actually determined by how
* long we wait before setting `refilling` to false and start the
* sending. Also, see `filter_process`.
*/
pw_properties_setf(filter_props, PW_KEY_NODE_FORCE_QUANTUM, "%u", impl->psamples);
pw_properties_setf(filter_props, PW_KEY_NODE_FORCE_RATE, "%u", impl->rate);
impl->ptp_sender = pw_filter_new(core, NULL, filter_props);
if (impl->ptp_sender == NULL)
return -errno;
pw_filter_add_listener(impl->ptp_sender, &impl->ptp_sender_listener,
&ptp_sender_events, impl);
n_params = 0;
params[n_params++] = spa_format_audio_raw_build(&b,
SPA_PARAM_EnumFormat, &impl->info.info.raw);
params[n_params++] = spa_format_audio_raw_build(&b,
SPA_PARAM_Format, &impl->info.info.raw);
ret = pw_filter_connect(impl->ptp_sender,
PW_FILTER_FLAG_RT_PROCESS,
params, n_params);
if (ret == 0) {
pw_log_info("created pw_filter for separate sender");
impl->separate_sender = true;
} else {
pw_log_error("failed to create pw_filter for separate sender");
impl->separate_sender = false;
}
return ret;
}
static int rtp_audio_init(struct impl *impl, struct pw_core *core, enum spa_direction direction, const char *ptp_driver)
{
if (direction == SPA_DIRECTION_INPUT)
impl->stream_events.process = rtp_audio_process_capture;
else
impl->stream_events.process = rtp_audio_process_playback;
impl->receive_rtp = rtp_audio_receive;
impl->stop_timer = rtp_audio_stop_timer;
impl->flush_timeout = rtp_audio_flush_timeout;
setup_ptp_sender(impl, core, direction, ptp_driver);
return 0;
}