Feature #13552 » 0002-thread_sync.c-rewrite-the-rest-using-using-ccan-list.patch
| thread_sync.c | ||
|---|---|---|
|
static VALUE rb_cMutex, rb_cQueue, rb_cSizedQueue, rb_cConditionVariable;
|
||
|
static VALUE rb_eClosedQueueError;
|
||
|
/* Mutex */
|
||
|
/* sync_waiter is always on-stack */
|
||
|
struct sync_waiter {
|
||
|
rb_thread_t *th;
|
||
| ... | ... | |
|
#define MUTEX_ALLOW_TRAP FL_USER1
|
||
|
static int
|
||
|
wakeup_one(struct list_head *head)
|
||
|
{
|
||
|
struct sync_waiter *cur = 0, *next = 0;
|
||
|
list_for_each_safe(head, cur, next, node) {
|
||
|
list_del_init(&cur->node);
|
||
|
if (cur->th->status != THREAD_KILLED) {
|
||
|
rb_threadptr_interrupt(cur->th);
|
||
|
cur->th->status = THREAD_RUNNABLE;
|
||
|
return TRUE;
|
||
|
}
|
||
|
}
|
||
|
return FALSE;
|
||
|
}
|
||
|
static void
|
||
|
wakeup_all(struct list_head *head)
|
||
|
{
|
||
|
struct sync_waiter *cur = 0, *next = 0;
|
||
|
list_for_each_safe(head, cur, next, node) {
|
||
|
list_del_init(&cur->node);
|
||
|
if (cur->th->status != THREAD_KILLED) {
|
||
|
rb_threadptr_interrupt(cur->th);
|
||
|
cur->th->status = THREAD_RUNNABLE;
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
/* Mutex */
|
||
|
typedef struct rb_mutex_struct {
|
||
|
struct rb_thread_struct volatile *th;
|
||
|
struct rb_mutex_struct *next_mutex;
|
||
| ... | ... | |
|
/* Queue */
|
||
|
enum {
|
||
|
QUEUE_QUE,
|
||
|
QUEUE_WAITERS,
|
||
|
SZQUEUE_WAITERS,
|
||
|
SZQUEUE_MAX,
|
||
|
END_QUEUE
|
||
|
};
|
||
|
PACKED_STRUCT_UNALIGNED(struct rb_queue {
|
||
|
struct list_head waitq;
|
||
|
VALUE que;
|
||
|
int num_waiting;
|
||
|
});
|
||
|
#define QUEUE_CLOSED FL_USER5
|
||
|
PACKED_STRUCT_UNALIGNED(struct rb_szqueue {
|
||
|
struct rb_queue q;
|
||
|
int num_waiting_push;
|
||
|
struct list_head pushq;
|
||
|
long max;
|
||
|
});
|
||
|
#define GET_QUEUE_QUE(q) get_array((q), QUEUE_QUE)
|
||
|
#define GET_QUEUE_WAITERS(q) get_array((q), QUEUE_WAITERS)
|
||
|
#define GET_SZQUEUE_WAITERS(q) get_array((q), SZQUEUE_WAITERS)
|
||
|
#define GET_SZQUEUE_MAX(q) RSTRUCT_GET((q), SZQUEUE_MAX)
|
||
|
#define GET_SZQUEUE_ULONGMAX(q) NUM2ULONG(GET_SZQUEUE_MAX(q))
|
||
|
static void
|
||
|
queue_mark(void *ptr)
|
||
|
{
|
||
|
struct rb_queue *q = ptr;
|
||
|
static VALUE
|
||
|
ary_buf_new(void)
|
||
|
/* no need to mark threads in waitq, they are on stack */
|
||
|
rb_gc_mark(q->que);
|
||
|
}
|
||
|
static size_t
|
||
|
queue_memsize(const void *ptr)
|
||
|
{
|
||
|
return rb_ary_tmp_new(1);
|
||
|
return sizeof(struct rb_queue);
|
||
|
}
|
||
|
static const rb_data_type_t queue_data_type = {
|
||
|
"queue",
|
||
|
{queue_mark, RUBY_TYPED_DEFAULT_FREE, queue_memsize,},
|
||
|
0, 0, RUBY_TYPED_FREE_IMMEDIATELY
|
||
|
};
|
||
|
static VALUE
|
||
|
get_array(VALUE obj, int idx)
|
||
|
queue_alloc(VALUE klass)
|
||
|
{
|
||
|
VALUE ary = RSTRUCT_GET(obj, idx);
|
||
|
if (!RB_TYPE_P(ary, T_ARRAY)) {
|
||
|
rb_raise(rb_eTypeError, "%+"PRIsVALUE" not initialized", obj);
|
||
|
}
|
||
|
return ary;
|
||
|
VALUE obj;
|
||
|
struct rb_queue *q;
|
||
|
obj = TypedData_Make_Struct(klass, struct rb_queue, &queue_data_type, q);
|
||
|
list_head_init(&q->waitq);
|
||
|
return obj;
|
||
|
}
|
||
|
static void
|
||
|
wakeup_first_thread(VALUE list)
|
||
|
static struct rb_queue *
|
||
|
queue_ptr(VALUE obj)
|
||
|
{
|
||
|
VALUE thread;
|
||
|
struct rb_queue *q;
|
||
|
while (!NIL_P(thread = rb_ary_shift(list))) {
|
||
|
if (RTEST(rb_thread_wakeup_alive(thread))) break;
|
||
|
}
|
||
|
TypedData_Get_Struct(obj, struct rb_queue, &queue_data_type, q);
|
||
|
return q;
|
||
|
}
|
||
|
#define QUEUE_CLOSED FL_USER5
|
||
|
static void
|
||
|
wakeup_all_threads(VALUE list)
|
||
|
szqueue_mark(void *ptr)
|
||
|
{
|
||
|
VALUE thread;
|
||
|
long i;
|
||
|
struct rb_szqueue *sq = ptr;
|
||
|
for (i=0; i<RARRAY_LEN(list); i++) {
|
||
|
thread = RARRAY_AREF(list, i);
|
||
|
rb_thread_wakeup_alive(thread);
|
||
|
}
|
||
|
rb_ary_clear(list);
|
||
|
queue_mark(&sq->q);
|
||
|
}
|
||
|
static size_t
|
||
|
szqueue_memsize(const void *ptr)
|
||
|
{
|
||
|
return sizeof(struct rb_szqueue);
|
||
|
}
|
||
|
static unsigned long
|
||
|
queue_length(VALUE self)
|
||
|
static const rb_data_type_t szqueue_data_type = {
|
||
|
"sized_queue",
|
||
|
{szqueue_mark, RUBY_TYPED_DEFAULT_FREE, szqueue_memsize,},
|
||
|
0, 0, RUBY_TYPED_FREE_IMMEDIATELY
|
||
|
};
|
||
|
static VALUE
|
||
|
szqueue_alloc(VALUE klass)
|
||
|
{
|
||
|
VALUE que = GET_QUEUE_QUE(self);
|
||
|
return RARRAY_LEN(que);
|
||
|
struct rb_szqueue *sq;
|
||
|
VALUE obj = TypedData_Make_Struct(klass, struct rb_szqueue,
|
||
|
&szqueue_data_type, sq);
|
||
|
list_head_init(&sq->q.waitq);
|
||
|
list_head_init(&sq->pushq);
|
||
|
return obj;
|
||
|
}
|
||
|
static unsigned long
|
||
|
queue_num_waiting(VALUE self)
|
||
|
static struct rb_szqueue *
|
||
|
szqueue_ptr(VALUE obj)
|
||
|
{
|
||
|
VALUE waiters = GET_QUEUE_WAITERS(self);
|
||
|
return RARRAY_LEN(waiters);
|
||
|
struct rb_szqueue *sq;
|
||
|
TypedData_Get_Struct(obj, struct rb_szqueue, &szqueue_data_type, sq);
|
||
|
return sq;
|
||
|
}
|
||
|
static unsigned long
|
||
|
szqueue_num_waiting_producer(VALUE self)
|
||
|
static VALUE
|
||
|
ary_buf_new(void)
|
||
|
{
|
||
|
VALUE waiters = GET_SZQUEUE_WAITERS(self);
|
||
|
return RARRAY_LEN(waiters);
|
||
|
return rb_ary_tmp_new(1);
|
||
|
}
|
||
|
static VALUE
|
||
|
check_array(VALUE obj, VALUE ary)
|
||
|
{
|
||
|
if (!RB_TYPE_P(ary, T_ARRAY)) {
|
||
|
rb_raise(rb_eTypeError, "%+"PRIsVALUE" not initialized", obj);
|
||
|
}
|
||
|
return ary;
|
||
|
}
|
||
|
static long
|
||
|
queue_length(VALUE self, struct rb_queue *q)
|
||
|
{
|
||
|
return RARRAY_LEN(check_array(self, q->que));
|
||
|
}
|
||
|
static int
|
||
| ... | ... | |
|
}
|
||
|
static VALUE
|
||
|
queue_closed_result(VALUE self)
|
||
|
queue_closed_result(VALUE self, struct rb_queue *q)
|
||
|
{
|
||
|
assert(queue_length(self) == 0);
|
||
|
assert(queue_length(self, q) == 0);
|
||
|
return Qnil;
|
||
|
}
|
||
|
static VALUE
|
||
|
queue_do_close(VALUE self, int is_szq)
|
||
|
{
|
||
|
if (!queue_closed_p(self)) {
|
||
|
FL_SET(self, QUEUE_CLOSED);
|
||
|
if (queue_num_waiting(self) > 0) {
|
||
|
VALUE waiters = GET_QUEUE_WAITERS(self);
|
||
|
wakeup_all_threads(waiters);
|
||
|
}
|
||
|
if (is_szq && szqueue_num_waiting_producer(self) > 0) {
|
||
|
VALUE waiters = GET_SZQUEUE_WAITERS(self);
|
||
|
wakeup_all_threads(waiters);
|
||
|
}
|
||
|
}
|
||
|
return self;
|
||
|
}
|
||
|
/*
|
||
|
* Document-class: Queue
|
||
|
*
|
||
| ... | ... | |
|
static VALUE
|
||
|
rb_queue_initialize(VALUE self)
|
||
|
{
|
||
|
RSTRUCT_SET(self, QUEUE_QUE, ary_buf_new());
|
||
|
RSTRUCT_SET(self, QUEUE_WAITERS, ary_buf_new());
|
||
|
struct rb_queue *q = queue_ptr(self);
|
||
|
q->que = ary_buf_new();
|
||
|
list_head_init(&q->waitq);
|
||
|
return self;
|
||
|
}
|
||
|
static VALUE
|
||
|
queue_do_push(VALUE self, VALUE obj)
|
||
|
queue_do_push(VALUE self, struct rb_queue *q, VALUE obj)
|
||
|
{
|
||
|
if (queue_closed_p(self)) {
|
||
|
raise_closed_queue_error(self);
|
||
|
}
|
||
|
rb_ary_push(GET_QUEUE_QUE(self), obj);
|
||
|
wakeup_first_thread(GET_QUEUE_WAITERS(self));
|
||
|
rb_ary_push(check_array(self, q->que), obj);
|
||
|
wakeup_one(&q->waitq);
|
||
|
return self;
|
||
|
}
|
||
| ... | ... | |
|
static VALUE
|
||
|
rb_queue_close(VALUE self)
|
||
|
{
|
||
|
return queue_do_close(self, FALSE);
|
||
|
struct rb_queue *q = queue_ptr(self);
|
||
|
if (!queue_closed_p(self)) {
|
||
|
FL_SET(self, QUEUE_CLOSED);
|
||
|
wakeup_all(&q->waitq);
|
||
|
}
|
||
|
return self;
|
||
|
}
|
||
|
/*
|
||
| ... | ... | |
|
static VALUE
|
||
|
rb_queue_push(VALUE self, VALUE obj)
|
||
|
{
|
||
|
return queue_do_push(self, obj);
|
||
|
return queue_do_push(self, queue_ptr(self), obj);
|
||
|
}
|
||
|
static VALUE
|
||
|
queue_sleep(VALUE arg)
|
||
|
{
|
||
|
rb_thread_sleep_deadly_allow_spurious_wakeup();
|
||
|
return Qnil;
|
||
|
}
|
||
|
struct waiting_delete {
|
||
|
VALUE waiting;
|
||
|
VALUE th;
|
||
|
struct queue_waiter {
|
||
|
struct sync_waiter w;
|
||
|
union {
|
||
|
struct rb_queue *q;
|
||
|
struct rb_szqueue *sq;
|
||
|
} as;
|
||
|
};
|
||
|
static VALUE
|
||
|
queue_delete_from_waiting(struct waiting_delete *p)
|
||
|
queue_sleep_done(VALUE p)
|
||
|
{
|
||
|
rb_ary_delete(p->waiting, p->th);
|
||
|
return Qnil;
|
||
|
struct queue_waiter *qw = (struct queue_waiter *)p;
|
||
|
list_del(&qw->w.node);
|
||
|
qw->as.q->num_waiting--;
|
||
|
return Qfalse;
|
||
|
}
|
||
|
static VALUE
|
||
|
queue_sleep(VALUE arg)
|
||
|
szqueue_sleep_done(VALUE p)
|
||
|
{
|
||
|
rb_thread_sleep_deadly_allow_spurious_wakeup();
|
||
|
return Qnil;
|
||
|
struct queue_waiter *qw = (struct queue_waiter *)p;
|
||
|
list_del(&qw->w.node);
|
||
|
qw->as.sq->num_waiting_push--;
|
||
|
return Qfalse;
|
||
|
}
|
||
|
static VALUE
|
||
|
queue_do_pop(VALUE self, int should_block)
|
||
|
queue_do_pop(VALUE self, struct rb_queue *q, int should_block)
|
||
|
{
|
||
|
struct waiting_delete args;
|
||
|
args.waiting = GET_QUEUE_WAITERS(self);
|
||
|
args.th = rb_thread_current();
|
||
|
check_array(self, q->que);
|
||
|
while (queue_length(self) == 0) {
|
||
|
while (RARRAY_LEN(q->que) == 0) {
|
||
|
if (!should_block) {
|
||
|
rb_raise(rb_eThreadError, "queue empty");
|
||
|
}
|
||
|
else if (queue_closed_p(self)) {
|
||
|
return queue_closed_result(self);
|
||
|
return queue_closed_result(self, q);
|
||
|
}
|
||
|
else {
|
||
|
assert(queue_length(self) == 0);
|
||
|
struct queue_waiter qw;
|
||
|
assert(RARRAY_LEN(q->que) == 0);
|
||
|
assert(queue_closed_p(self) == 0);
|
||
|
rb_ary_push(args.waiting, args.th);
|
||
|
rb_ensure(queue_sleep, Qfalse, queue_delete_from_waiting, (VALUE)&args);
|
||
|
qw.w.th = GET_THREAD();
|
||
|
qw.as.q = q;
|
||
|
list_add_tail(&qw.as.q->waitq, &qw.w.node);
|
||
|
qw.as.q->num_waiting++;
|
||
|
rb_ensure(queue_sleep, Qfalse, queue_sleep_done, (VALUE)&qw);
|
||
|
}
|
||
|
}
|
||
|
return rb_ary_shift(GET_QUEUE_QUE(self));
|
||
|
return rb_ary_shift(q->que);
|
||
|
}
|
||
|
static int
|
||
| ... | ... | |
|
rb_queue_pop(int argc, VALUE *argv, VALUE self)
|
||
|
{
|
||
|
int should_block = queue_pop_should_block(argc, argv);
|
||
|
return queue_do_pop(self, should_block);
|
||
|
return queue_do_pop(self, queue_ptr(self), should_block);
|
||
|
}
|
||
|
/*
|
||
| ... | ... | |
|
static VALUE
|
||
|
rb_queue_empty_p(VALUE self)
|
||
|
{
|
||
|
return queue_length(self) == 0 ? Qtrue : Qfalse;
|
||
|
return queue_length(self, queue_ptr(self)) == 0 ? Qtrue : Qfalse;
|
||
|
}
|
||
|
/*
|
||
| ... | ... | |
|
static VALUE
|
||
|
rb_queue_clear(VALUE self)
|
||
|
{
|
||
|
rb_ary_clear(GET_QUEUE_QUE(self));
|
||
|
struct rb_queue *q = queue_ptr(self);
|
||
|
rb_ary_clear(check_array(self, q->que));
|
||
|
return self;
|
||
|
}
|
||
| ... | ... | |
|
static VALUE
|
||
|
rb_queue_length(VALUE self)
|
||
|
{
|
||
|
unsigned long len = queue_length(self);
|
||
|
return ULONG2NUM(len);
|
||
|
return LONG2NUM(queue_length(self, queue_ptr(self)));
|
||
|
}
|
||
|
/*
|
||
| ... | ... | |
|
static VALUE
|
||
|
rb_queue_num_waiting(VALUE self)
|
||
|
{
|
||
|
unsigned long len = queue_num_waiting(self);
|
||
|
return ULONG2NUM(len);
|
||
|
struct rb_queue *q = queue_ptr(self);
|
||
|
return INT2NUM(q->num_waiting);
|
||
|
}
|
||
|
/*
|
||
| ... | ... | |
|
rb_szqueue_initialize(VALUE self, VALUE vmax)
|
||
|
{
|
||
|
long max;
|
||
|
struct rb_szqueue *sq = szqueue_ptr(self);
|
||
|
max = NUM2LONG(vmax);
|
||
|
if (max <= 0) {
|
||
|
rb_raise(rb_eArgError, "queue size must be positive");
|
||
|
}
|
||
|
RSTRUCT_SET(self, QUEUE_QUE, ary_buf_new());
|
||
|
RSTRUCT_SET(self, QUEUE_WAITERS, ary_buf_new());
|
||
|
RSTRUCT_SET(self, SZQUEUE_WAITERS, ary_buf_new());
|
||
|
RSTRUCT_SET(self, SZQUEUE_MAX, vmax);
|
||
|
sq->q.que = ary_buf_new();
|
||
|
list_head_init(&sq->q.waitq);
|
||
|
list_head_init(&sq->pushq);
|
||
|
sq->max = max;
|
||
|
return self;
|
||
|
}
|
||
| ... | ... | |
|
static VALUE
|
||
|
rb_szqueue_close(VALUE self)
|
||
|
{
|
||
|
return queue_do_close(self, TRUE);
|
||
|
if (!queue_closed_p(self)) {
|
||
|
struct rb_szqueue *sq = szqueue_ptr(self);
|
||
|
FL_SET(self, QUEUE_CLOSED);
|
||
|
wakeup_all(&sq->q.waitq);
|
||
|
wakeup_all(&sq->pushq);
|
||
|
}
|
||
|
return self;
|
||
|
}
|
||
|
/*
|
||
| ... | ... | |
|
static VALUE
|
||
|
rb_szqueue_max_get(VALUE self)
|
||
|
{
|
||
|
return GET_SZQUEUE_MAX(self);
|
||
|
return LONG2NUM(szqueue_ptr(self)->max);
|
||
|
}
|
||
|
/*
|
||
| ... | ... | |
|
static VALUE
|
||
|
rb_szqueue_max_set(VALUE self, VALUE vmax)
|
||
|
{
|
||
|
long max = NUM2LONG(vmax), diff = 0;
|
||
|
VALUE t;
|
||
|
long max = NUM2LONG(vmax);
|
||
|
long diff = 0;
|
||
|
struct rb_szqueue *sq = szqueue_ptr(self);
|
||
|
if (max <= 0) {
|
||
|
rb_raise(rb_eArgError, "queue size must be positive");
|
||
|
}
|
||
|
if ((unsigned long)max > GET_SZQUEUE_ULONGMAX(self)) {
|
||
|
diff = max - GET_SZQUEUE_ULONGMAX(self);
|
||
|
if (max > sq->max) {
|
||
|
diff = max - sq->max;
|
||
|
}
|
||
|
RSTRUCT_SET(self, SZQUEUE_MAX, vmax);
|
||
|
while (diff-- > 0 && !NIL_P(t = rb_ary_shift(GET_SZQUEUE_WAITERS(self)))) {
|
||
|
rb_thread_wakeup_alive(t);
|
||
|
sq->max = max;
|
||
|
while (diff-- > 0 && wakeup_one(&sq->pushq)) {
|
||
|
/* keep waking more up */
|
||
|
}
|
||
|
return vmax;
|
||
|
}
|
||
| ... | ... | |
|
static VALUE
|
||
|
rb_szqueue_push(int argc, VALUE *argv, VALUE self)
|
||
|
{
|
||
|
struct waiting_delete args;
|
||
|
struct rb_szqueue *sq = szqueue_ptr(self);
|
||
|
int should_block = szqueue_push_should_block(argc, argv);
|
||
|
args.waiting = GET_SZQUEUE_WAITERS(self);
|
||
|
args.th = rb_thread_current();
|
||
|
while (queue_length(self) >= GET_SZQUEUE_ULONGMAX(self)) {
|
||
|
while (queue_length(self, &sq->q) >= sq->max) {
|
||
|
if (!should_block) {
|
||
|
rb_raise(rb_eThreadError, "queue full");
|
||
|
}
|
||
| ... | ... | |
|
goto closed;
|
||
|
}
|
||
|
else {
|
||
|
rb_ary_push(args.waiting, args.th);
|
||
|
rb_ensure(queue_sleep, Qfalse, queue_delete_from_waiting, (VALUE)&args);
|
||
|
struct queue_waiter qw;
|
||
|
qw.w.th = GET_THREAD();
|
||
|
qw.as.sq = sq;
|
||
|
list_add_tail(&sq->pushq, &qw.w.node);
|
||
|
sq->num_waiting_push++;
|
||
|
rb_ensure(queue_sleep, Qfalse, szqueue_sleep_done, (VALUE)&qw);
|
||
|
}
|
||
|
}
|
||
| ... | ... | |
|
raise_closed_queue_error(self);
|
||
|
}
|
||
|
return queue_do_push(self, argv[0]);
|
||
|
return queue_do_push(self, &sq->q, argv[0]);
|
||
|
}
|
||
|
static VALUE
|
||
|
szqueue_do_pop(VALUE self, int should_block)
|
||
|
{
|
||
|
VALUE retval = queue_do_pop(self, should_block);
|
||
|
struct rb_szqueue *sq = szqueue_ptr(self);
|
||
|
VALUE retval = queue_do_pop(self, &sq->q, should_block);
|
||
|
if (queue_length(self) < GET_SZQUEUE_ULONGMAX(self)) {
|
||
|
wakeup_first_thread(GET_SZQUEUE_WAITERS(self));
|
||
|
if (queue_length(self, &sq->q) < sq->max) {
|
||
|
wakeup_one(&sq->pushq);
|
||
|
}
|
||
|
return retval;
|
||
| ... | ... | |
|
static VALUE
|
||
|
rb_szqueue_clear(VALUE self)
|
||
|
{
|
||
|
rb_ary_clear(GET_QUEUE_QUE(self));
|
||
|
wakeup_all_threads(GET_SZQUEUE_WAITERS(self));
|
||
|
struct rb_szqueue *sq = szqueue_ptr(self);
|
||
|
rb_ary_clear(check_array(self, sq->q.que));
|
||
|
wakeup_all(&sq->pushq);
|
||
|
return self;
|
||
|
}
|
||
|
static VALUE
|
||
|
rb_szqueue_length(VALUE self)
|
||
|
{
|
||
|
struct rb_szqueue *sq = szqueue_ptr(self);
|
||
|
return LONG2NUM(queue_length(self, &sq->q));
|
||
|
}
|
||
|
/*
|
||
|
* Document-method: SizedQueue#num_waiting
|
||
|
*
|
||
| ... | ... | |
|
static VALUE
|
||
|
rb_szqueue_num_waiting(VALUE self)
|
||
|
{
|
||
|
long len = queue_num_waiting(self) + szqueue_num_waiting_producer(self);
|
||
|
return ULONG2NUM(len);
|
||
|
struct rb_szqueue *sq = szqueue_ptr(self);
|
||
|
return INT2NUM(sq->q.num_waiting + sq->num_waiting_push);
|
||
|
}
|
||
|
/* ConditionalVariable */
|
||
|
/*
|
||
|
* Document-method: SizedQueue#empty?
|
||
|
* call-seq: empty?
|
||
|
*
|
||
|
* Returns +true+ if the queue is empty.
|
||
|
*/
|
||
|
static VALUE
|
||
|
rb_szqueue_empty_p(VALUE self)
|
||
|
{
|
||
|
struct rb_szqueue *sq = szqueue_ptr(self);
|
||
|
return queue_length(self, &sq->q) == 0 ? Qtrue : Qfalse;
|
||
|
}
|
||
|
enum {
|
||
|
CONDVAR_WAITERS,
|
||
|
END_CONDVAR
|
||
|
};
|
||
|
#define GET_CONDVAR_WAITERS(cv) get_array((cv), CONDVAR_WAITERS)
|
||
|
/* ConditionalVariable */
|
||
|
/* TODO: maybe this can be IMEMO */
|
||
|
struct rb_condvar {
|
||
|
struct list_head waitq;
|
||
|
};
|
||
|
/*
|
||
|
* Document-class: ConditionVariable
|
||
| ... | ... | |
|
* }
|
||
|
*/
|
||
|
static size_t
|
||
|
condvar_memsize(const void *ptr)
|
||
|
{
|
||
|
return sizeof(struct rb_condvar);
|
||
|
}
|
||
|
static const rb_data_type_t cv_data_type = {
|
||
|
"condvar",
|
||
|
{0, RUBY_TYPED_DEFAULT_FREE, condvar_memsize,},
|
||
|
0, 0, RUBY_TYPED_FREE_IMMEDIATELY
|
||
|
};
|
||
|
static struct rb_condvar *
|
||
|
condvar_ptr(VALUE self)
|
||
|
{
|
||
|
struct rb_condvar *cv;
|
||
|
TypedData_Get_Struct(self, struct rb_condvar, &cv_data_type, cv);
|
||
|
return cv;
|
||
|
}
|
||
|
static VALUE
|
||
|
condvar_alloc(VALUE klass)
|
||
|
{
|
||
|
struct rb_condvar *cv;
|
||
|
VALUE obj;
|
||
|
obj = TypedData_Make_Struct(klass, struct rb_condvar, &cv_data_type, cv);
|
||
|
list_head_init(&cv->waitq);
|
||
|
return obj;
|
||
|
}
|
||
|
/*
|
||
|
* Document-method: ConditionVariable::new
|
||
|
*
|
||
| ... | ... | |
|
static VALUE
|
||
|
rb_condvar_initialize(VALUE self)
|
||
|
{
|
||
|
RSTRUCT_SET(self, CONDVAR_WAITERS, ary_buf_new());
|
||
|
struct rb_condvar *cv = condvar_ptr(self);;
|
||
|
list_head_init(&cv->waitq);
|
||
|
return self;
|
||
|
}
|
||
| ... | ... | |
|
}
|
||
|
static VALUE
|
||
|
delete_current_thread(VALUE ary)
|
||
|
delete_from_waitq(struct sync_waiter *w)
|
||
|
{
|
||
|
return rb_ary_delete(ary, rb_thread_current());
|
||
|
list_del(&w->node);
|
||
|
return Qnil;
|
||
|
}
|
||
|
/*
|
||
| ... | ... | |
|
static VALUE
|
||
|
rb_condvar_wait(int argc, VALUE *argv, VALUE self)
|
||
|
{
|
||
|
VALUE waiters = GET_CONDVAR_WAITERS(self);
|
||
|
struct rb_condvar *cv = condvar_ptr(self);
|
||
|
VALUE mutex, timeout;
|
||
|
struct sleep_call args;
|
||
|
struct sync_waiter w;
|
||
|
rb_scan_args(argc, argv, "11", &mutex, &timeout);
|
||
|
args.mutex = mutex;
|
||
|
args.timeout = timeout;
|
||
|
rb_ary_push(waiters, rb_thread_current());
|
||
|
rb_ensure(do_sleep, (VALUE)&args, delete_current_thread, waiters);
|
||
|
w.th = GET_THREAD();
|
||
|
list_add_tail(&cv->waitq, &w.node);
|
||
|
rb_ensure(do_sleep, (VALUE)&args, delete_from_waitq, (VALUE)&w);
|
||
|
return self;
|
||
|
}
|
||
| ... | ... | |
|
static VALUE
|
||
|
rb_condvar_signal(VALUE self)
|
||
|
{
|
||
|
wakeup_first_thread(GET_CONDVAR_WAITERS(self));
|
||
|
struct rb_condvar *cv = condvar_ptr(self);
|
||
|
wakeup_one(&cv->waitq);
|
||
|
return self;
|
||
|
}
|
||
| ... | ... | |
|
static VALUE
|
||
|
rb_condvar_broadcast(VALUE self)
|
||
|
{
|
||
|
wakeup_all_threads(GET_CONDVAR_WAITERS(self));
|
||
|
struct rb_condvar *cv = condvar_ptr(self);
|
||
|
wakeup_all(&cv->waitq);
|
||
|
return self;
|
||
|
}
|
||
| ... | ... | |
|
rb_define_method(rb_cMutex, "owned?", rb_mutex_owned_p, 0);
|
||
|
/* Queue */
|
||
|
rb_cQueue = rb_struct_define_without_accessor_under(
|
||
|
rb_cThread,
|
||
|
"Queue", rb_cObject, rb_struct_alloc_noinit,
|
||
|
"que", "waiters", NULL);
|
||
|
rb_cQueue = rb_define_class_under(rb_cThread, "Queue", rb_cObject);
|
||
|
rb_define_alloc_func(rb_cQueue, queue_alloc);
|
||
|
rb_eClosedQueueError = rb_define_class("ClosedQueueError", rb_eStopIteration);
|
||
| ... | ... | |
|
rb_define_alias(rb_cQueue, "shift", "pop");
|
||
|
rb_define_alias(rb_cQueue, "size", "length");
|
||
|
rb_cSizedQueue = rb_struct_define_without_accessor_under(
|
||
|
rb_cThread,
|
||
|
"SizedQueue", rb_cQueue, rb_struct_alloc_noinit,
|
||
|
"que", "waiters", "queue_waiters", "size", NULL);
|
||
|
rb_cSizedQueue = rb_define_class_under(rb_cThread, "SizedQueue", rb_cQueue);
|
||
|
rb_define_alloc_func(rb_cSizedQueue, szqueue_alloc);
|
||
|
rb_define_method(rb_cSizedQueue, "initialize", rb_szqueue_initialize, 1);
|
||
|
rb_define_method(rb_cSizedQueue, "close", rb_szqueue_close, 0);
|
||
| ... | ... | |
|
rb_define_method(rb_cSizedQueue, "max=", rb_szqueue_max_set, 1);
|
||
|
rb_define_method(rb_cSizedQueue, "push", rb_szqueue_push, -1);
|
||
|
rb_define_method(rb_cSizedQueue, "pop", rb_szqueue_pop, -1);
|
||
|
rb_define_method(rb_cSizedQueue, "empty?", rb_szqueue_empty_p, 0);
|
||
|
rb_define_method(rb_cSizedQueue, "clear", rb_szqueue_clear, 0);
|
||
|
rb_define_method(rb_cSizedQueue, "length", rb_szqueue_length, 0);
|
||
|
rb_define_method(rb_cSizedQueue, "num_waiting", rb_szqueue_num_waiting, 0);
|
||
|
rb_define_alias(rb_cSizedQueue, "enq", "push");
|
||
|
rb_define_alias(rb_cSizedQueue, "<<", "push");
|
||
|
rb_define_alias(rb_cSizedQueue, "deq", "pop");
|
||
|
rb_define_alias(rb_cSizedQueue, "shift", "pop");
|
||
|
rb_define_alias(rb_cSizedQueue, "size", "length");
|
||
|
/* CVar */
|
||
|
rb_cConditionVariable = rb_struct_define_without_accessor_under(
|
||
|
rb_cThread,
|
||
|
"ConditionVariable", rb_cObject, rb_struct_alloc_noinit,
|
||
|
"waiters", NULL);
|
||
|
rb_cConditionVariable = rb_define_class_under(rb_cThread,
|
||
|
"ConditionVariable", rb_cObject);
|
||
|
rb_define_alloc_func(rb_cConditionVariable, condvar_alloc);
|
||
|
id_sleep = rb_intern("sleep");
|
||
|
-
|
||
- « Previous
- 1
- 2
- Next »