pulseaudio/src/pulsecore/flist.c

231 lines
5.3 KiB
C

/***
This file is part of PulseAudio.
Copyright 2006-2008 Lennart Poettering
PulseAudio is free software; you can redistribute it and/or modify
it under the terms of the GNU Lesser General Public License as
published by the Free Software Foundation; either version 2.1 of the
License, or (at your option) any later version.
PulseAudio is distributed in the hope that it will be useful, but
WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
Lesser General Public License for more details.
You should have received a copy of the GNU Lesser General Public
License along with PulseAudio; if not, write to the Free Software
Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307
USA.
***/
#ifdef HAVE_CONFIG_H
#include <config.h>
#endif
#include <pulse/xmalloc.h>
#include <pulsecore/atomic.h>
#include <pulsecore/log.h>
#include <pulsecore/thread.h>
#include <pulsecore/macro.h>
#include <pulsecore/core-util.h>
#include <pulsecore/macro.h>
#include "flist.h"
/* Algorithm is not perfect, in a few corner cases it will fail to pop
* from the flist although it isn't empty, and fail to push into the
* flist, although it isn't full.
*
* We keep a fixed size array of entries, each item is an atomic
* pointer.
*
* To accelerate finding of used/unused cells we maintain a read and a
* write index which is used like a ring buffer. After each push we
* increase the write index and after each pop we increase the read
* index.
*
* The indexes are incremented atomically and are never truncated to
* the buffer size. Instead we assume that the buffer size is a power
* of two and that the truncation can thus be done by applying a
* simple AND on read.
*
* To make sure that we do not look for empty cells indefinitely we
* maintain a length value which stores the number of used cells. From
* this value the number of unused cells is easily calculated. Please
* note that the length value is not updated atomically with the read
* and write index and might thus be a few cells off the real
* value. To deal with this we always look for N_EXTRA_SCAN extra
* cells when pushing/popping entries.
*
* It might make sense to replace this implementation with a link list
* stack or queue, which however requires DCAS to be simple. Patches
* welcome.
*
* Please note that this algorithm is home grown.*/
#define FLIST_SIZE 128
#define N_EXTRA_SCAN 3
/* For debugging purposes we can define _Y to put and extra thread
* yield between each operation. */
#ifdef PROFILE
#define _Y pa_thread_yield()
#else
#define _Y do { } while(0)
#endif
struct pa_flist {
unsigned size;
pa_atomic_t length;
pa_atomic_t read_idx;
pa_atomic_t write_idx;
};
#define PA_FLIST_CELLS(x) ((pa_atomic_ptr_t*) ((uint8_t*) (x) + PA_ALIGN(sizeof(struct pa_flist))))
pa_flist *pa_flist_new(unsigned size) {
pa_flist *l;
if (!size)
size = FLIST_SIZE;
pa_assert(pa_is_power_of_two(size));
l = pa_xmalloc0(PA_ALIGN(sizeof(pa_flist)) + (sizeof(pa_atomic_ptr_t) * size));
l->size = size;
pa_atomic_store(&l->read_idx, 0);
pa_atomic_store(&l->write_idx, 0);
pa_atomic_store(&l->length, 0);
return l;
}
static unsigned reduce(pa_flist *l, unsigned value) {
return value & (l->size - 1);
}
void pa_flist_free(pa_flist *l, pa_free_cb_t free_cb) {
pa_assert(l);
if (free_cb) {
pa_atomic_ptr_t*cells;
unsigned idx;
cells = PA_FLIST_CELLS(l);
for (idx = 0; idx < l->size; idx ++) {
void *p;
if ((p = pa_atomic_ptr_load(&cells[idx])))
free_cb(p);
}
}
pa_xfree(l);
}
int pa_flist_push(pa_flist*l, void *p) {
unsigned idx, n;
pa_atomic_ptr_t*cells;
#ifdef PROFILE
unsigned len;
#endif
pa_assert(l);
pa_assert(p);
cells = PA_FLIST_CELLS(l);
n = l->size + N_EXTRA_SCAN - (unsigned) pa_atomic_load(&l->length);
#ifdef PROFILE
len = n;
#endif
_Y;
idx = reduce(l, (unsigned) pa_atomic_load(&l->write_idx));
for (; n > 0 ; n--) {
_Y;
if (pa_atomic_ptr_cmpxchg(&cells[idx], NULL, p)) {
_Y;
pa_atomic_inc(&l->write_idx);
_Y;
pa_atomic_inc(&l->length);
return 0;
}
_Y;
idx = reduce(l, idx + 1);
}
#ifdef PROFILE
if (len > N_EXTRA_SCAN)
pa_log_warn("Didn't find free cell after %u iterations.", len);
#endif
return -1;
}
void* pa_flist_pop(pa_flist*l) {
unsigned idx, n;
pa_atomic_ptr_t *cells;
#ifdef PROFILE
unsigned len;
#endif
pa_assert(l);
cells = PA_FLIST_CELLS(l);
n = (unsigned) pa_atomic_load(&l->length) + N_EXTRA_SCAN;
#ifdef PROFILE
len = n;
#endif
_Y;
idx = reduce(l, (unsigned) pa_atomic_load(&l->read_idx));
for (; n > 0 ; n--) {
void *p;
_Y;
p = pa_atomic_ptr_load(&cells[idx]);
if (p) {
_Y;
if (!pa_atomic_ptr_cmpxchg(&cells[idx], p, NULL))
continue;
_Y;
pa_atomic_inc(&l->read_idx);
_Y;
pa_atomic_dec(&l->length);
return p;
}
_Y;
idx = reduce(l, idx+1);
}
#ifdef PROFILE
if (len > N_EXTRA_SCAN)
pa_log_warn("Didn't find used cell after %u iterations.", len);
#endif
return NULL;
}