summaryrefslogtreecommitdiff
path: root/thread_sync.c
diff options
context:
space:
mode:
Diffstat (limited to 'thread_sync.c')
-rw-r--r--thread_sync.c1175
1 files changed, 391 insertions, 784 deletions
diff --git a/thread_sync.c b/thread_sync.c
index c0a0ca7103..e3916c97cb 100644
--- a/thread_sync.c
+++ b/thread_sync.c
@@ -2,12 +2,12 @@
#include "ccan/list/list.h"
#include "builtin.h"
-static VALUE rb_cMutex, rb_cQueue, rb_cSizedQueue, rb_cConditionVariable;
-static VALUE rb_eClosedQueueError;
+static VALUE rb_cMutex, rb_eClosedQueueError;
/* Mutex */
typedef struct rb_mutex_struct {
- rb_fiber_t *fiber;
+ rb_serial_t ec_serial;
+ rb_thread_t *th; // even if the fiber is collected, we might need access to the thread in mutex_free
struct rb_mutex_struct *next_mutex;
struct ccan_list_head waitq; /* protected by GVL */
} rb_mutex_t;
@@ -80,33 +80,7 @@ static void rb_mutex_abandon_all(rb_mutex_t *mutexes);
static void rb_mutex_abandon_keeping_mutexes(rb_thread_t *th);
static void rb_mutex_abandon_locking_mutex(rb_thread_t *th);
#endif
-static const char* rb_mutex_unlock_th(rb_mutex_t *mutex, rb_thread_t *th, rb_fiber_t *fiber);
-
-/*
- * Document-class: Thread::Mutex
- *
- * Thread::Mutex implements a simple semaphore that can be used to
- * coordinate access to shared data from multiple concurrent threads.
- *
- * Example:
- *
- * semaphore = Thread::Mutex.new
- *
- * a = Thread.new {
- * semaphore.synchronize {
- * # access shared resource
- * }
- * }
- *
- * b = Thread.new {
- * semaphore.synchronize {
- * # access shared resource
- * }
- * }
- *
- */
-
-#define mutex_mark ((void(*)(void*))0)
+static const char* rb_mutex_unlock_th(rb_mutex_t *mutex, rb_thread_t *th, rb_serial_t ec_serial);
static size_t
rb_mutex_num_waiting(rb_mutex_t *mutex)
@@ -123,13 +97,18 @@ rb_mutex_num_waiting(rb_mutex_t *mutex)
rb_thread_t* rb_fiber_threadptr(const rb_fiber_t *fiber);
+static bool
+mutex_locked_p(rb_mutex_t *mutex)
+{
+ return mutex->ec_serial != 0;
+}
+
static void
mutex_free(void *ptr)
{
rb_mutex_t *mutex = ptr;
- if (mutex->fiber) {
- /* rb_warn("free locked mutex"); */
- const char *err = rb_mutex_unlock_th(mutex, rb_fiber_threadptr(mutex->fiber), mutex->fiber);
+ if (mutex_locked_p(mutex)) {
+ const char *err = rb_mutex_unlock_th(mutex, mutex->th, 0);
if (err) rb_bug("%s", err);
}
ruby_xfree(ptr);
@@ -143,8 +122,8 @@ mutex_memsize(const void *ptr)
static const rb_data_type_t mutex_data_type = {
"mutex",
- {mutex_mark, mutex_free, mutex_memsize,},
- 0, 0, RUBY_TYPED_WB_PROTECTED | RUBY_TYPED_FREE_IMMEDIATELY
+ {NULL, mutex_free, mutex_memsize,},
+ 0, 0, RUBY_TYPED_FREE_IMMEDIATELY
};
static rb_mutex_t *
@@ -175,41 +154,24 @@ mutex_alloc(VALUE klass)
return obj;
}
-/*
- * call-seq:
- * Thread::Mutex.new -> mutex
- *
- * Creates a new Mutex
- */
-static VALUE
-mutex_initialize(VALUE self)
-{
- return self;
-}
-
VALUE
rb_mutex_new(void)
{
return mutex_alloc(rb_cMutex);
}
-/*
- * call-seq:
- * mutex.locked? -> true or false
- *
- * Returns +true+ if this lock is currently held by some thread.
- */
VALUE
rb_mutex_locked_p(VALUE self)
{
rb_mutex_t *mutex = mutex_ptr(self);
- return RBOOL(mutex->fiber);
+ return RBOOL(mutex_locked_p(mutex));
}
static void
thread_mutex_insert(rb_thread_t *thread, rb_mutex_t *mutex)
{
+ RUBY_ASSERT(!mutex->next_mutex);
if (thread->keeping_mutexes) {
mutex->next_mutex = thread->keeping_mutexes;
}
@@ -234,45 +196,50 @@ thread_mutex_remove(rb_thread_t *thread, rb_mutex_t *mutex)
}
static void
-mutex_locked(rb_thread_t *th, VALUE self)
+mutex_set_owner(rb_mutex_t *mutex, rb_thread_t *th, rb_serial_t ec_serial)
{
- rb_mutex_t *mutex = mutex_ptr(self);
+ mutex->th = th;
+ mutex->ec_serial = ec_serial;
+}
+static void
+mutex_locked(rb_mutex_t *mutex, rb_thread_t *th, rb_serial_t ec_serial)
+{
+ mutex_set_owner(mutex, th, ec_serial);
thread_mutex_insert(th, mutex);
}
-/*
- * call-seq:
- * mutex.try_lock -> true or false
- *
- * Attempts to obtain the lock and returns immediately. Returns +true+ if the
- * lock was granted.
- */
-VALUE
-rb_mutex_trylock(VALUE self)
+static inline bool
+do_mutex_trylock(rb_mutex_t *mutex, rb_thread_t *th, rb_serial_t ec_serial)
{
- rb_mutex_t *mutex = mutex_ptr(self);
-
- if (mutex->fiber == 0) {
+ if (mutex->ec_serial == 0) {
RUBY_DEBUG_LOG("%p ok", mutex);
- rb_fiber_t *fiber = GET_EC()->fiber_ptr;
- rb_thread_t *th = GET_THREAD();
- mutex->fiber = fiber;
-
- mutex_locked(th, self);
- return Qtrue;
+ mutex_locked(mutex, th, ec_serial);
+ return true;
}
else {
RUBY_DEBUG_LOG("%p ng", mutex);
- return Qfalse;
+ return false;
}
}
static VALUE
-mutex_owned_p(rb_fiber_t *fiber, rb_mutex_t *mutex)
+rb_mut_trylock(rb_execution_context_t *ec, VALUE self)
+{
+ return RBOOL(do_mutex_trylock(mutex_ptr(self), ec->thread_ptr, rb_ec_serial(ec)));
+}
+
+VALUE
+rb_mutex_trylock(VALUE self)
{
- return RBOOL(mutex->fiber == fiber);
+ return rb_mut_trylock(GET_EC(), self);
+}
+
+static VALUE
+mutex_owned_p(rb_serial_t ec_serial, rb_mutex_t *mutex)
+{
+ return RBOOL(mutex->ec_serial == ec_serial);
}
static VALUE
@@ -292,13 +259,29 @@ delete_from_waitq(VALUE value)
static inline rb_atomic_t threadptr_get_interrupts(rb_thread_t *th);
+struct mutex_args {
+ VALUE self;
+ rb_mutex_t *mutex;
+ rb_execution_context_t *ec;
+};
+
+static inline void
+mutex_args_init(struct mutex_args *args, VALUE mutex)
+{
+ args->self = mutex;
+ args->mutex = mutex_ptr(mutex);
+ args->ec = GET_EC();
+}
+
static VALUE
-do_mutex_lock(VALUE self, int interruptible_p)
+do_mutex_lock(struct mutex_args *args, int interruptible_p)
{
- rb_execution_context_t *ec = GET_EC();
+ VALUE self = args->self;
+ rb_execution_context_t *ec = args->ec;
rb_thread_t *th = ec->thread_ptr;
rb_fiber_t *fiber = ec->fiber_ptr;
- rb_mutex_t *mutex = mutex_ptr(self);
+ rb_serial_t ec_serial = rb_ec_serial(ec);
+ rb_mutex_t *mutex = args->mutex;
rb_atomic_t saved_ints = 0;
/* When running trap handler */
@@ -307,13 +290,13 @@ do_mutex_lock(VALUE self, int interruptible_p)
rb_raise(rb_eThreadError, "can't be called from trap context");
}
- if (rb_mutex_trylock(self) == Qfalse) {
- if (mutex->fiber == fiber) {
+ if (!do_mutex_trylock(mutex, th, ec_serial)) {
+ if (mutex->ec_serial == ec_serial) {
rb_raise(rb_eThreadError, "deadlock; recursive locking");
}
- while (mutex->fiber != fiber) {
- VM_ASSERT(mutex->fiber != NULL);
+ while (mutex->ec_serial != ec_serial) {
+ VM_ASSERT(mutex->ec_serial != 0);
VALUE scheduler = rb_fiber_scheduler_current();
if (scheduler != Qnil) {
@@ -327,12 +310,12 @@ do_mutex_lock(VALUE self, int interruptible_p)
rb_ensure(call_rb_fiber_scheduler_block, self, delete_from_waitq, (VALUE)&sync_waiter);
- if (!mutex->fiber) {
- mutex->fiber = fiber;
+ if (!mutex->ec_serial) {
+ mutex_set_owner(mutex, th, ec_serial);
}
}
else {
- if (!th->vm->thread_ignore_deadlock && rb_fiber_threadptr(mutex->fiber) == th) {
+ if (!th->vm->thread_ignore_deadlock && mutex->th == th) {
rb_raise(rb_eThreadError, "deadlock; lock already owned by another fiber belonging to the same thread");
}
@@ -358,6 +341,7 @@ do_mutex_lock(VALUE self, int interruptible_p)
rb_ractor_sleeper_threads_inc(th->ractor);
rb_check_deadlock(th->ractor);
+ RUBY_ASSERT(!th->locking_mutex);
th->locking_mutex = self;
ccan_list_add_tail(&mutex->waitq, &sync_waiter.node);
@@ -367,14 +351,13 @@ do_mutex_lock(VALUE self, int interruptible_p)
ccan_list_del(&sync_waiter.node);
// unlocked by another thread while sleeping
- if (!mutex->fiber) {
- mutex->fiber = fiber;
+ if (!mutex->ec_serial) {
+ mutex_set_owner(mutex, th, ec_serial);
}
rb_ractor_sleeper_threads_dec(th->ractor);
th->status = prev_status;
th->locking_mutex = Qfalse;
- th->locking_mutex = Qfalse;
RUBY_DEBUG_LOG("%p wakeup", mutex);
}
@@ -382,10 +365,13 @@ do_mutex_lock(VALUE self, int interruptible_p)
if (interruptible_p) {
/* release mutex before checking for interrupts...as interrupt checking
* code might call rb_raise() */
- if (mutex->fiber == fiber) mutex->fiber = 0;
+ if (mutex->ec_serial == ec_serial) {
+ mutex->th = NULL;
+ mutex->ec_serial = 0;
+ }
RUBY_VM_CHECK_INTS_BLOCKING(th->ec); /* may release mutex */
- if (!mutex->fiber) {
- mutex->fiber = fiber;
+ if (!mutex->ec_serial) {
+ mutex_set_owner(mutex, th, ec_serial);
}
}
else {
@@ -404,13 +390,13 @@ do_mutex_lock(VALUE self, int interruptible_p)
}
if (saved_ints) th->ec->interrupt_flag = saved_ints;
- if (mutex->fiber == fiber) mutex_locked(th, self);
+ if (mutex->ec_serial == ec_serial) mutex_locked(mutex, th, ec_serial);
}
RUBY_DEBUG_LOG("%p locked", mutex);
// assertion
- if (mutex_owned_p(fiber, mutex) == Qfalse) rb_bug("do_mutex_lock: mutex is not owned.");
+ if (mutex_owned_p(ec_serial, mutex) == Qfalse) rb_bug("do_mutex_lock: mutex is not owned.");
return self;
}
@@ -418,52 +404,57 @@ do_mutex_lock(VALUE self, int interruptible_p)
static VALUE
mutex_lock_uninterruptible(VALUE self)
{
- return do_mutex_lock(self, 0);
+ struct mutex_args args;
+ mutex_args_init(&args, self);
+ return do_mutex_lock(&args, 0);
+}
+
+static VALUE
+rb_mut_lock(rb_execution_context_t *ec, VALUE self)
+{
+ struct mutex_args args = {
+ .self = self,
+ .mutex = mutex_ptr(self),
+ .ec = ec,
+ };
+ return do_mutex_lock(&args, 1);
}
-/*
- * call-seq:
- * mutex.lock -> self
- *
- * Attempts to grab the lock and waits if it isn't available.
- * Raises +ThreadError+ if +mutex+ was locked by the current thread.
- */
VALUE
rb_mutex_lock(VALUE self)
{
- return do_mutex_lock(self, 1);
+ struct mutex_args args;
+ mutex_args_init(&args, self);
+ return do_mutex_lock(&args, 1);
+}
+
+static VALUE
+rb_mut_owned_p(rb_execution_context_t *ec, VALUE self)
+{
+ return mutex_owned_p(rb_ec_serial(ec), mutex_ptr(self));
}
-/*
- * call-seq:
- * mutex.owned? -> true or false
- *
- * Returns +true+ if this lock is currently held by current thread.
- */
VALUE
rb_mutex_owned_p(VALUE self)
{
- rb_fiber_t *fiber = GET_EC()->fiber_ptr;
- rb_mutex_t *mutex = mutex_ptr(self);
-
- return mutex_owned_p(fiber, mutex);
+ return rb_mut_owned_p(GET_EC(), self);
}
static const char *
-rb_mutex_unlock_th(rb_mutex_t *mutex, rb_thread_t *th, rb_fiber_t *fiber)
+rb_mutex_unlock_th(rb_mutex_t *mutex, rb_thread_t *th, rb_serial_t ec_serial)
{
RUBY_DEBUG_LOG("%p", mutex);
- if (mutex->fiber == 0) {
+ if (mutex->ec_serial == 0) {
return "Attempt to unlock a mutex which is not locked";
}
- else if (mutex->fiber != fiber) {
+ else if (ec_serial && mutex->ec_serial != ec_serial) {
return "Attempt to unlock a mutex which is locked by another thread/fiber";
}
struct sync_waiter *cur = 0, *next;
- mutex->fiber = 0;
+ mutex->ec_serial = 0;
thread_mutex_remove(th, mutex);
ccan_list_for_each_safe(&mutex->waitq, cur, next, node) {
@@ -494,6 +485,24 @@ rb_mutex_unlock_th(rb_mutex_t *mutex, rb_thread_t *th, rb_fiber_t *fiber)
return NULL;
}
+static void
+do_mutex_unlock(struct mutex_args *args)
+{
+ const char *err;
+ rb_mutex_t *mutex = args->mutex;
+ rb_thread_t *th = rb_ec_thread_ptr(args->ec);
+
+ err = rb_mutex_unlock_th(mutex, th, rb_ec_serial(args->ec));
+ if (err) rb_raise(rb_eThreadError, "%s", err);
+}
+
+static VALUE
+do_mutex_unlock_safe(VALUE args)
+{
+ do_mutex_unlock((struct mutex_args *)args);
+ return Qnil;
+}
+
/*
* call-seq:
* mutex.unlock -> self
@@ -504,13 +513,21 @@ rb_mutex_unlock_th(rb_mutex_t *mutex, rb_thread_t *th, rb_fiber_t *fiber)
VALUE
rb_mutex_unlock(VALUE self)
{
- const char *err;
- rb_mutex_t *mutex = mutex_ptr(self);
- rb_thread_t *th = GET_THREAD();
-
- err = rb_mutex_unlock_th(mutex, th, GET_EC()->fiber_ptr);
- if (err) rb_raise(rb_eThreadError, "%s", err);
+ struct mutex_args args;
+ mutex_args_init(&args, self);
+ do_mutex_unlock(&args);
+ return self;
+}
+static VALUE
+rb_mut_unlock(rb_execution_context_t *ec, VALUE self)
+{
+ struct mutex_args args = {
+ .self = self,
+ .mutex = mutex_ptr(self),
+ .ec = ec,
+ };
+ do_mutex_unlock(&args);
return self;
}
@@ -541,117 +558,99 @@ rb_mutex_abandon_all(rb_mutex_t *mutexes)
while (mutexes) {
mutex = mutexes;
mutexes = mutex->next_mutex;
- mutex->fiber = 0;
+ mutex->ec_serial = 0;
mutex->next_mutex = 0;
ccan_list_head_init(&mutex->waitq);
}
}
#endif
-static VALUE
-rb_mutex_sleep_forever(VALUE self)
-{
- rb_thread_sleep_deadly_allow_spurious_wakeup(self, Qnil, 0);
- return Qnil;
-}
+struct rb_mutex_sleep_arguments {
+ VALUE self;
+ VALUE timeout;
+};
static VALUE
-rb_mutex_wait_for(VALUE time)
-{
- rb_hrtime_t *rel = (rb_hrtime_t *)time;
- /* permit spurious check */
- return RBOOL(sleep_hrtime(GET_THREAD(), *rel, 0));
-}
-
-VALUE
-rb_mutex_sleep(VALUE self, VALUE timeout)
+mutex_sleep_begin(VALUE _arguments)
{
- struct timeval t;
+ struct rb_mutex_sleep_arguments *arguments = (struct rb_mutex_sleep_arguments *)_arguments;
+ VALUE timeout = arguments->timeout;
VALUE woken = Qtrue;
- if (!NIL_P(timeout)) {
- t = rb_time_interval(timeout);
- }
-
- rb_mutex_unlock(self);
- time_t beg = time(0);
-
VALUE scheduler = rb_fiber_scheduler_current();
if (scheduler != Qnil) {
rb_fiber_scheduler_kernel_sleep(scheduler, timeout);
- mutex_lock_uninterruptible(self);
}
else {
if (NIL_P(timeout)) {
- rb_ensure(rb_mutex_sleep_forever, self, mutex_lock_uninterruptible, self);
+ rb_thread_sleep_deadly_allow_spurious_wakeup(arguments->self, Qnil, 0);
}
else {
- rb_hrtime_t rel = rb_timeval2hrtime(&t);
- woken = rb_ensure(rb_mutex_wait_for, (VALUE)&rel, mutex_lock_uninterruptible, self);
+ struct timeval timeout_value = rb_time_interval(timeout);
+ rb_hrtime_t relative_timeout = rb_timeval2hrtime(&timeout_value);
+ /* permit spurious check */
+ woken = RBOOL(sleep_hrtime(GET_THREAD(), relative_timeout, 0));
}
}
- RUBY_VM_CHECK_INTS_BLOCKING(GET_EC());
- if (!woken) return Qnil;
- time_t end = time(0) - beg;
- return TIMET2NUM(end);
+ return woken;
}
-/*
- * call-seq:
- * mutex.sleep(timeout = nil) -> number or nil
- *
- * Releases the lock and sleeps +timeout+ seconds if it is given and
- * non-nil or forever. Raises +ThreadError+ if +mutex+ wasn't locked by
- * the current thread.
- *
- * When the thread is next woken up, it will attempt to reacquire
- * the lock.
- *
- * Note that this method can wakeup without explicit Thread#wakeup call.
- * For example, receiving signal and so on.
- *
- * Returns the slept time in seconds if woken up, or +nil+ if timed out.
- */
static VALUE
-mutex_sleep(int argc, VALUE *argv, VALUE self)
+rb_mut_sleep(rb_execution_context_t *ec, VALUE self, VALUE timeout)
{
- VALUE timeout;
+ if (!NIL_P(timeout)) {
+ // Validate the argument:
+ rb_time_interval(timeout);
+ }
+
+ rb_mut_unlock(ec, self);
+ time_t beg = time(0);
- timeout = rb_check_arity(argc, 0, 1) ? argv[0] : Qnil;
- return rb_mutex_sleep(self, timeout);
+ struct rb_mutex_sleep_arguments arguments = {
+ .self = self,
+ .timeout = timeout,
+ };
+
+ VALUE woken = rb_ec_ensure(ec, mutex_sleep_begin, (VALUE)&arguments, mutex_lock_uninterruptible, self);
+
+ RUBY_VM_CHECK_INTS_BLOCKING(ec);
+ if (!woken) return Qnil;
+ time_t end = time(0) - beg;
+ return TIMET2NUM(end);
}
-/*
- * call-seq:
- * mutex.synchronize { ... } -> result of the block
- *
- * Obtains a lock, runs the block, and releases the lock when the block
- * completes. See the example under Thread::Mutex.
- */
+VALUE
+rb_mutex_sleep(VALUE self, VALUE timeout)
+{
+ return rb_mut_sleep(GET_EC(), self, timeout);
+}
VALUE
-rb_mutex_synchronize(VALUE mutex, VALUE (*func)(VALUE arg), VALUE arg)
+rb_mutex_synchronize(VALUE self, VALUE (*func)(VALUE arg), VALUE arg)
{
- rb_mutex_lock(mutex);
- return rb_ensure(func, arg, rb_mutex_unlock, mutex);
+ struct mutex_args args;
+ mutex_args_init(&args, self);
+ do_mutex_lock(&args, 1);
+ return rb_ec_ensure(args.ec, func, arg, do_mutex_unlock_safe, (VALUE)&args);
}
-/*
- * call-seq:
- * mutex.synchronize { ... } -> result of the block
- *
- * Obtains a lock, runs the block, and releases the lock when the block
- * completes. See the example under Thread::Mutex.
- */
static VALUE
-rb_mutex_synchronize_m(VALUE self)
+do_ec_yield(VALUE _ec)
{
- if (!rb_block_given_p()) {
- rb_raise(rb_eThreadError, "must be called with a block");
- }
+ return rb_ec_yield((rb_execution_context_t *)_ec, Qundef);
+}
- return rb_mutex_synchronize(self, rb_yield, Qundef);
+VALUE
+rb_mut_synchronize(rb_execution_context_t *ec, VALUE self)
+{
+ struct mutex_args args = {
+ .self = self,
+ .mutex = mutex_ptr(self),
+ .ec = ec,
+ };
+ do_mutex_lock(&args, 1);
+ return rb_ec_ensure(args.ec, do_ec_yield, (VALUE)ec, do_mutex_unlock_safe, (VALUE)&args);
}
void
@@ -667,46 +666,61 @@ rb_mutex_allow_trap(VALUE self, int val)
/* Queue */
-#define queue_waitq(q) UNALIGNED_MEMBER_PTR(q, waitq)
-#define queue_list(q) UNALIGNED_MEMBER_PTR(q, que)
-RBIMPL_ATTR_PACKED_STRUCT_UNALIGNED_BEGIN()
struct rb_queue {
struct ccan_list_head waitq;
rb_serial_t fork_gen;
- const VALUE que;
+ long capa;
+ long len;
+ long offset;
+ VALUE *buffer;
int num_waiting;
-} RBIMPL_ATTR_PACKED_STRUCT_UNALIGNED_END();
+};
+
+#define szqueue_waitq(sq) &sq->q.waitq
+#define szqueue_pushq(sq) &sq->pushq
-#define szqueue_waitq(sq) UNALIGNED_MEMBER_PTR(sq, q.waitq)
-#define szqueue_list(sq) UNALIGNED_MEMBER_PTR(sq, q.que)
-#define szqueue_pushq(sq) UNALIGNED_MEMBER_PTR(sq, pushq)
-RBIMPL_ATTR_PACKED_STRUCT_UNALIGNED_BEGIN()
struct rb_szqueue {
struct rb_queue q;
int num_waiting_push;
struct ccan_list_head pushq;
long max;
-} RBIMPL_ATTR_PACKED_STRUCT_UNALIGNED_END();
+};
static void
-queue_mark(void *ptr)
+queue_mark_and_move(void *ptr)
{
struct rb_queue *q = ptr;
-
/* no need to mark threads in waitq, they are on stack */
- rb_gc_mark(q->que);
+ for (long index = 0; index < q->len; index++) {
+ rb_gc_mark_and_move(&q->buffer[((q->offset + index) % q->capa)]);
+ }
+}
+
+static void
+queue_free(void *ptr)
+{
+ struct rb_queue *q = ptr;
+ if (q->buffer) {
+ ruby_sized_xfree(q->buffer, q->capa * sizeof(VALUE));
+ }
}
static size_t
queue_memsize(const void *ptr)
{
- return sizeof(struct rb_queue);
+ const struct rb_queue *q = ptr;
+ return sizeof(struct rb_queue) + (q->capa * sizeof(VALUE));
}
static const rb_data_type_t queue_data_type = {
- "queue",
- {queue_mark, RUBY_TYPED_DEFAULT_FREE, queue_memsize,},
- 0, 0, RUBY_TYPED_FREE_IMMEDIATELY|RUBY_TYPED_WB_PROTECTED
+ .wrap_struct_name = "Thread::Queue",
+ .function = {
+ .dmark = queue_mark_and_move,
+ .dfree = queue_free,
+ .dsize = queue_memsize,
+ .dcompact = queue_mark_and_move,
+ },
+ .flags = RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_WB_PROTECTED,
};
static VALUE
@@ -716,27 +730,27 @@ queue_alloc(VALUE klass)
struct rb_queue *q;
obj = TypedData_Make_Struct(klass, struct rb_queue, &queue_data_type, q);
- ccan_list_head_init(queue_waitq(q));
+ ccan_list_head_init(&q->waitq);
return obj;
}
-static int
+static inline bool
queue_fork_check(struct rb_queue *q)
{
rb_serial_t fork_gen = GET_VM()->fork_gen;
- if (q->fork_gen == fork_gen) {
- return 0;
+ if (RB_LIKELY(q->fork_gen == fork_gen)) {
+ return false;
}
/* forked children can't reach into parent thread stacks */
q->fork_gen = fork_gen;
- ccan_list_head_init(queue_waitq(q));
+ ccan_list_head_init(&q->waitq);
q->num_waiting = 0;
- return 1;
+ return true;
}
-static struct rb_queue *
-queue_ptr(VALUE obj)
+static inline struct rb_queue *
+raw_queue_ptr(VALUE obj)
{
struct rb_queue *q;
@@ -746,6 +760,22 @@ queue_ptr(VALUE obj)
return q;
}
+static inline void
+check_queue(VALUE obj, struct rb_queue *q)
+{
+ if (RB_UNLIKELY(q->buffer == NULL)) {
+ rb_raise(rb_eTypeError, "%+"PRIsVALUE" not initialized", obj);
+ }
+}
+
+static inline struct rb_queue *
+queue_ptr(VALUE obj)
+{
+ struct rb_queue *q = raw_queue_ptr(obj);
+ check_queue(obj, q);
+ return q;
+}
+
#define QUEUE_CLOSED FL_USER5
static rb_hrtime_t
@@ -765,23 +795,37 @@ queue_timeout2hrtime(VALUE timeout)
}
static void
-szqueue_mark(void *ptr)
+szqueue_mark_and_move(void *ptr)
{
struct rb_szqueue *sq = ptr;
- queue_mark(&sq->q);
+ queue_mark_and_move(&sq->q);
+}
+
+static void
+szqueue_free(void *ptr)
+{
+ struct rb_szqueue *sq = ptr;
+ queue_free(&sq->q);
}
static size_t
szqueue_memsize(const void *ptr)
{
- return sizeof(struct rb_szqueue);
+ const struct rb_szqueue *sq = ptr;
+ return sizeof(struct rb_szqueue) + (sq->q.capa * sizeof(VALUE));
}
static const rb_data_type_t szqueue_data_type = {
- "sized_queue",
- {szqueue_mark, RUBY_TYPED_DEFAULT_FREE, szqueue_memsize,},
- 0, 0, RUBY_TYPED_FREE_IMMEDIATELY|RUBY_TYPED_WB_PROTECTED
+ .wrap_struct_name = "Thread::SizedQueue",
+ .function = {
+ .dmark = szqueue_mark_and_move,
+ .dfree = szqueue_free,
+ .dsize = szqueue_memsize,
+ .dcompact = szqueue_mark_and_move,
+ },
+ .parent = &queue_data_type,
+ .flags = RUBY_TYPED_FREE_IMMEDIATELY | RUBY_TYPED_WB_PROTECTED,
};
static VALUE
@@ -795,13 +839,13 @@ szqueue_alloc(VALUE klass)
return obj;
}
-static struct rb_szqueue *
-szqueue_ptr(VALUE obj)
+static inline struct rb_szqueue *
+raw_szqueue_ptr(VALUE obj)
{
struct rb_szqueue *sq;
TypedData_Get_Struct(obj, struct rb_szqueue, &szqueue_data_type, sq);
- if (queue_fork_check(&sq->q)) {
+ if (RB_UNLIKELY(queue_fork_check(&sq->q))) {
ccan_list_head_init(szqueue_pushq(sq));
sq->num_waiting_push = 0;
}
@@ -809,28 +853,15 @@ szqueue_ptr(VALUE obj)
return sq;
}
-static VALUE
-ary_buf_new(void)
-{
- return rb_ary_hidden_new(1);
-}
-
-static VALUE
-check_array(VALUE obj, VALUE ary)
-{
- if (!RB_TYPE_P(ary, T_ARRAY)) {
- rb_raise(rb_eTypeError, "%+"PRIsVALUE" not initialized", obj);
- }
- return ary;
-}
-
-static long
-queue_length(VALUE self, struct rb_queue *q)
+static inline struct rb_szqueue *
+szqueue_ptr(VALUE obj)
{
- return RARRAY_LEN(check_array(self, q->que));
+ struct rb_szqueue *sq = raw_szqueue_ptr(obj);
+ check_queue(obj, &sq->q);
+ return sq;
}
-static int
+static inline bool
queue_closed_p(VALUE self)
{
return FL_TEST_RAW(self, QUEUE_CLOSED) != 0;
@@ -854,171 +885,95 @@ raise_closed_queue_error(VALUE self)
static VALUE
queue_closed_result(VALUE self, struct rb_queue *q)
{
- assert(queue_length(self, q) == 0);
+ RUBY_ASSERT(q->len == 0);
return Qnil;
}
-/*
- * Document-class: Thread::Queue
- *
- * The Thread::Queue class implements multi-producer, multi-consumer
- * queues. It is especially useful in threaded programming when
- * information must be exchanged safely between multiple threads. The
- * Thread::Queue class implements all the required locking semantics.
- *
- * The class implements FIFO (first in, first out) type of queue.
- * In a FIFO queue, the first tasks added are the first retrieved.
- *
- * Example:
- *
- * queue = Thread::Queue.new
- *
- * producer = Thread.new do
- * 5.times do |i|
- * sleep rand(i) # simulate expense
- * queue << i
- * puts "#{i} produced"
- * end
- * end
- *
- * consumer = Thread.new do
- * 5.times do |i|
- * value = queue.pop
- * sleep rand(i/2) # simulate expense
- * puts "consumed #{value}"
- * end
- * end
- *
- * consumer.join
- *
- */
+#define QUEUE_INITIAL_CAPA 8
-/*
- * Document-method: Queue::new
- *
- * call-seq:
- * Thread::Queue.new -> empty_queue
- * Thread::Queue.new(enumerable) -> queue
- *
- * Creates a new queue instance, optionally using the contents of an +enumerable+
- * for its initial state.
- *
- * Example:
- *
- * q = Thread::Queue.new
- * #=> #<Thread::Queue:0x00007ff7501110d0>
- * q.empty?
- * #=> true
- *
- * q = Thread::Queue.new([1, 2, 3])
- * #=> #<Thread::Queue:0x00007ff7500ec500>
- * q.empty?
- * #=> false
- * q.pop
- * #=> 1
- */
+static inline void
+ring_buffer_init(struct rb_queue *q, long initial_capa)
+{
+ q->buffer = ALLOC_N(VALUE, initial_capa);
+ q->capa = initial_capa;
+}
-static VALUE
-rb_queue_initialize(int argc, VALUE *argv, VALUE self)
+static inline void
+ring_buffer_expand(struct rb_queue *q)
{
- VALUE initial;
- struct rb_queue *q = queue_ptr(self);
- if ((argc = rb_scan_args(argc, argv, "01", &initial)) == 1) {
- initial = rb_to_array(initial);
- }
- RB_OBJ_WRITE(self, queue_list(q), ary_buf_new());
- ccan_list_head_init(queue_waitq(q));
- if (argc == 1) {
- rb_ary_concat(q->que, initial);
- }
- return self;
+ RUBY_ASSERT(q->capa > 0);
+ VALUE *new_buffer = ALLOC_N(VALUE, q->capa * 2);
+ MEMCPY(new_buffer, q->buffer + q->offset, VALUE, q->capa - q->offset);
+ MEMCPY(new_buffer + (q->capa - q->offset), q->buffer, VALUE, q->offset);
+ VALUE *old_buffer = q->buffer;
+ q->buffer = new_buffer;
+ q->offset = 0;
+ ruby_sized_xfree(old_buffer, q->capa * sizeof(VALUE));
+ q->capa *= 2;
}
-static VALUE
-queue_do_push(VALUE self, struct rb_queue *q, VALUE obj)
+static void
+ring_buffer_push(VALUE self, struct rb_queue *q, VALUE obj)
{
- if (queue_closed_p(self)) {
- raise_closed_queue_error(self);
+ if (RB_UNLIKELY(q->len >= q->capa)) {
+ ring_buffer_expand(q);
}
- rb_ary_push(check_array(self, q->que), obj);
- wakeup_one(queue_waitq(q));
- return self;
+ RUBY_ASSERT(q->capa > q->len);
+ long index = (q->offset + q->len) % q->capa;
+ q->len++;
+ RB_OBJ_WRITE(self, &q->buffer[index], obj);
}
-/*
- * Document-method: Thread::Queue#close
- * call-seq:
- * close
- *
- * Closes the queue. A closed queue cannot be re-opened.
- *
- * After the call to close completes, the following are true:
- *
- * - +closed?+ will return true
- *
- * - +close+ will be ignored.
- *
- * - calling enq/push/<< will raise a +ClosedQueueError+.
- *
- * - when +empty?+ is false, calling deq/pop/shift will return an object
- * from the queue as usual.
- * - when +empty?+ is true, deq(false) will not suspend the thread and will return nil.
- * deq(true) will raise a +ThreadError+.
- *
- * ClosedQueueError is inherited from StopIteration, so that you can break loop block.
- *
- * Example:
- *
- * q = Thread::Queue.new
- * Thread.new{
- * while e = q.deq # wait for nil to break loop
- * # ...
- * end
- * }
- * q.close
- */
-
static VALUE
-rb_queue_close(VALUE self)
+ring_buffer_shift(struct rb_queue *q)
{
- struct rb_queue *q = queue_ptr(self);
-
- if (!queue_closed_p(self)) {
- FL_SET(self, QUEUE_CLOSED);
-
- wakeup_all(queue_waitq(q));
+ if (!q->len) {
+ return Qnil;
}
- return self;
+ VALUE obj = q->buffer[q->offset];
+ q->len--;
+ if (q->len == 0) {
+ q->offset = 0;
+ }
+ else {
+ q->offset = (q->offset + 1) % q->capa;
+ }
+ return obj;
}
-/*
- * Document-method: Thread::Queue#closed?
- * call-seq: closed?
- *
- * Returns +true+ if the queue is closed.
- */
-
static VALUE
-rb_queue_closed_p(VALUE self)
+queue_initialize(rb_execution_context_t *ec, VALUE self, VALUE initial)
{
- return RBOOL(queue_closed_p(self));
+ struct rb_queue *q = raw_queue_ptr(self);
+ ccan_list_head_init(&q->waitq);
+ if (NIL_P(initial)) {
+ ring_buffer_init(q, QUEUE_INITIAL_CAPA);
+ }
+ else {
+ initial = rb_to_array(initial);
+ long len = RARRAY_LEN(initial);
+ long initial_capa = QUEUE_INITIAL_CAPA;
+ while (initial_capa < len) {
+ initial_capa *= 2;
+ }
+ ring_buffer_init(q, initial_capa);
+ MEMCPY(q->buffer, RARRAY_CONST_PTR(initial), VALUE, len);
+ q->len = len;
+ }
+ return self;
}
-/*
- * Document-method: Thread::Queue#push
- * call-seq:
- * push(object)
- * enq(object)
- * <<(object)
- *
- * Pushes the given +object+ to the queue.
- */
-
static VALUE
-rb_queue_push(VALUE self, VALUE obj)
+queue_do_push(VALUE self, struct rb_queue *q, VALUE obj)
{
- return queue_do_push(self, queue_ptr(self), obj);
+ check_queue(self, q);
+ if (queue_closed_p(self)) {
+ raise_closed_queue_error(self);
+ }
+ ring_buffer_push(self, q, obj);
+ wakeup_one(&q->waitq);
+ return self;
}
static VALUE
@@ -1059,12 +1014,11 @@ szqueue_sleep_done(VALUE p)
return Qfalse;
}
-static VALUE
-queue_do_pop(VALUE self, struct rb_queue *q, int should_block, VALUE timeout)
+static inline VALUE
+queue_do_pop(rb_execution_context_t *ec, VALUE self, struct rb_queue *q, VALUE non_block, VALUE timeout)
{
- check_array(self, q->que);
- if (RARRAY_LEN(q->que) == 0) {
- if (!should_block) {
+ if (q->len == 0) {
+ if (RTEST(non_block)) {
rb_raise(rb_eThreadError, "queue empty");
}
@@ -1074,22 +1028,20 @@ queue_do_pop(VALUE self, struct rb_queue *q, int should_block, VALUE timeout)
}
rb_hrtime_t end = queue_timeout2hrtime(timeout);
- while (RARRAY_LEN(q->que) == 0) {
+ while (q->len == 0) {
if (queue_closed_p(self)) {
return queue_closed_result(self, q);
}
else {
- rb_execution_context_t *ec = GET_EC();
-
- assert(RARRAY_LEN(q->que) == 0);
- assert(queue_closed_p(self) == 0);
+ RUBY_ASSERT(q->len == 0);
+ RUBY_ASSERT(queue_closed_p(self) == 0);
struct queue_waiter queue_waiter = {
.w = {.self = self, .th = ec->thread_ptr, .fiber = nonblocking_fiber(ec->fiber_ptr)},
.as = {.q = q}
};
- struct ccan_list_head *waitq = queue_waitq(q);
+ struct ccan_list_head *waitq = &q->waitq;
ccan_list_add_tail(waitq, &queue_waiter.w.node);
queue_waiter.as.q->num_waiting++;
@@ -1106,116 +1058,32 @@ queue_do_pop(VALUE self, struct rb_queue *q, int should_block, VALUE timeout)
}
}
- return rb_ary_shift(q->que);
+ return ring_buffer_shift(q);
}
static VALUE
rb_queue_pop(rb_execution_context_t *ec, VALUE self, VALUE non_block, VALUE timeout)
{
- return queue_do_pop(self, queue_ptr(self), !RTEST(non_block), timeout);
+ return queue_do_pop(ec, self, queue_ptr(self), non_block, timeout);
}
-/*
- * Document-method: Thread::Queue#empty?
- * call-seq: empty?
- *
- * Returns +true+ if the queue is empty.
- */
-
-static VALUE
-rb_queue_empty_p(VALUE self)
-{
- return RBOOL(queue_length(self, queue_ptr(self)) == 0);
-}
-
-/*
- * Document-method: Thread::Queue#clear
- *
- * Removes all objects from the queue.
- */
-
-static VALUE
-rb_queue_clear(VALUE self)
-{
- struct rb_queue *q = queue_ptr(self);
-
- rb_ary_clear(check_array(self, q->que));
- return self;
-}
-
-/*
- * Document-method: Thread::Queue#length
- * call-seq:
- * length
- * size
- *
- * Returns the length of the queue.
- */
-
-static VALUE
-rb_queue_length(VALUE self)
+static void
+queue_clear(struct rb_queue *q)
{
- return LONG2NUM(queue_length(self, queue_ptr(self)));
+ q->len = 0;
+ q->offset = 0;
}
-NORETURN(static VALUE rb_queue_freeze(VALUE self));
-/*
- * call-seq:
- * freeze
- *
- * The queue can't be frozen, so this method raises an exception:
- * Thread::Queue.new.freeze # Raises TypeError (cannot freeze #<Thread::Queue:0x...>)
- *
- */
static VALUE
-rb_queue_freeze(VALUE self)
+szqueue_initialize(rb_execution_context_t *ec, VALUE self, VALUE vmax)
{
- rb_raise(rb_eTypeError, "cannot freeze " "%+"PRIsVALUE, self);
- UNREACHABLE_RETURN(self);
-}
-
-/*
- * Document-method: Thread::Queue#num_waiting
- *
- * Returns the number of threads waiting on the queue.
- */
-
-static VALUE
-rb_queue_num_waiting(VALUE self)
-{
- struct rb_queue *q = queue_ptr(self);
-
- return INT2NUM(q->num_waiting);
-}
-
-/*
- * Document-class: Thread::SizedQueue
- *
- * This class represents queues of specified size capacity. The push operation
- * may be blocked if the capacity is full.
- *
- * See Thread::Queue for an example of how a Thread::SizedQueue works.
- */
-
-/*
- * Document-method: SizedQueue::new
- * call-seq: new(max)
- *
- * Creates a fixed-length queue with a maximum size of +max+.
- */
-
-static VALUE
-rb_szqueue_initialize(VALUE self, VALUE vmax)
-{
- long max;
- struct rb_szqueue *sq = szqueue_ptr(self);
+ long max = NUM2LONG(vmax);
+ struct rb_szqueue *sq = raw_szqueue_ptr(self);
- max = NUM2LONG(vmax);
if (max <= 0) {
rb_raise(rb_eArgError, "queue size must be positive");
}
-
- RB_OBJ_WRITE(self, szqueue_list(sq), ary_buf_new());
+ ring_buffer_init(&sq->q, QUEUE_INITIAL_CAPA);
ccan_list_head_init(szqueue_waitq(sq));
ccan_list_head_init(szqueue_pushq(sq));
sq->max = max;
@@ -1223,74 +1091,12 @@ rb_szqueue_initialize(VALUE self, VALUE vmax)
return self;
}
-/*
- * Document-method: Thread::SizedQueue#close
- * call-seq:
- * close
- *
- * Similar to Thread::Queue#close.
- *
- * The difference is behavior with waiting enqueuing threads.
- *
- * If there are waiting enqueuing threads, they are interrupted by
- * raising ClosedQueueError('queue closed').
- */
-static VALUE
-rb_szqueue_close(VALUE self)
-{
- if (!queue_closed_p(self)) {
- struct rb_szqueue *sq = szqueue_ptr(self);
-
- FL_SET(self, QUEUE_CLOSED);
- wakeup_all(szqueue_waitq(sq));
- wakeup_all(szqueue_pushq(sq));
- }
- return self;
-}
-
-/*
- * Document-method: Thread::SizedQueue#max
- *
- * Returns the maximum size of the queue.
- */
-
-static VALUE
-rb_szqueue_max_get(VALUE self)
-{
- return LONG2NUM(szqueue_ptr(self)->max);
-}
-
-/*
- * Document-method: Thread::SizedQueue#max=
- * call-seq: max=(number)
- *
- * Sets the maximum size of the queue to the given +number+.
- */
-
-static VALUE
-rb_szqueue_max_set(VALUE self, VALUE vmax)
-{
- long max = NUM2LONG(vmax);
- long diff = 0;
- struct rb_szqueue *sq = szqueue_ptr(self);
-
- if (max <= 0) {
- rb_raise(rb_eArgError, "queue size must be positive");
- }
- if (max > sq->max) {
- diff = max - sq->max;
- }
- sq->max = max;
- sync_wakeup(szqueue_pushq(sq), diff);
- return vmax;
-}
-
static VALUE
rb_szqueue_push(rb_execution_context_t *ec, VALUE self, VALUE object, VALUE non_block, VALUE timeout)
{
struct rb_szqueue *sq = szqueue_ptr(self);
- if (queue_length(self, &sq->q) >= sq->max) {
+ if (sq->q.len >= sq->max) {
if (RTEST(non_block)) {
rb_raise(rb_eThreadError, "queue full");
}
@@ -1301,12 +1107,11 @@ rb_szqueue_push(rb_execution_context_t *ec, VALUE self, VALUE object, VALUE non_
}
rb_hrtime_t end = queue_timeout2hrtime(timeout);
- while (queue_length(self, &sq->q) >= sq->max) {
+ while (sq->q.len >= sq->max) {
if (queue_closed_p(self)) {
raise_closed_queue_error(self);
}
else {
- rb_execution_context_t *ec = GET_EC();
struct queue_waiter queue_waiter = {
.w = {.self = self, .th = ec->thread_ptr, .fiber = nonblocking_fiber(ec->fiber_ptr)},
.as = {.sq = sq}
@@ -1333,85 +1138,17 @@ rb_szqueue_push(rb_execution_context_t *ec, VALUE self, VALUE object, VALUE non_
}
static VALUE
-szqueue_do_pop(VALUE self, int should_block, VALUE timeout)
+rb_szqueue_pop(rb_execution_context_t *ec, VALUE self, VALUE non_block, VALUE timeout)
{
struct rb_szqueue *sq = szqueue_ptr(self);
- VALUE retval = queue_do_pop(self, &sq->q, should_block, timeout);
+ VALUE retval = queue_do_pop(ec, self, &sq->q, non_block, timeout);
- if (queue_length(self, &sq->q) < sq->max) {
+ if (sq->q.len < sq->max) {
wakeup_one(szqueue_pushq(sq));
}
return retval;
}
-static VALUE
-rb_szqueue_pop(rb_execution_context_t *ec, VALUE self, VALUE non_block, VALUE timeout)
-{
- return szqueue_do_pop(self, !RTEST(non_block), timeout);
-}
-
-/*
- * Document-method: Thread::SizedQueue#clear
- *
- * Removes all objects from the queue.
- */
-
-static VALUE
-rb_szqueue_clear(VALUE self)
-{
- struct rb_szqueue *sq = szqueue_ptr(self);
-
- rb_ary_clear(check_array(self, sq->q.que));
- wakeup_all(szqueue_pushq(sq));
- return self;
-}
-
-/*
- * Document-method: Thread::SizedQueue#length
- * call-seq:
- * length
- * size
- *
- * Returns the length of the queue.
- */
-
-static VALUE
-rb_szqueue_length(VALUE self)
-{
- struct rb_szqueue *sq = szqueue_ptr(self);
-
- return LONG2NUM(queue_length(self, &sq->q));
-}
-
-/*
- * Document-method: Thread::SizedQueue#num_waiting
- *
- * Returns the number of threads waiting on the queue.
- */
-
-static VALUE
-rb_szqueue_num_waiting(VALUE self)
-{
- struct rb_szqueue *sq = szqueue_ptr(self);
-
- return INT2NUM(sq->q.num_waiting + sq->num_waiting_push);
-}
-
-/*
- * Document-method: Thread::SizedQueue#empty?
- * call-seq: empty?
- *
- * Returns +true+ if the queue is empty.
- */
-
-static VALUE
-rb_szqueue_empty_p(VALUE self)
-{
- struct rb_szqueue *sq = szqueue_ptr(self);
-
- return RBOOL(queue_length(self, &sq->q) == 0);
-}
-
/* ConditionalVariable */
struct rb_condvar {
@@ -1419,34 +1156,6 @@ struct rb_condvar {
rb_serial_t fork_gen;
};
-/*
- * Document-class: Thread::ConditionVariable
- *
- * ConditionVariable objects augment class Mutex. Using condition variables,
- * it is possible to suspend while in the middle of a critical section until a
- * resource becomes available.
- *
- * Example:
- *
- * mutex = Thread::Mutex.new
- * resource = Thread::ConditionVariable.new
- *
- * a = Thread.new {
- * mutex.synchronize {
- * # Thread 'a' now needs the resource
- * resource.wait(mutex)
- * # 'a' can now have the resource
- * }
- * }
- *
- * b = Thread.new {
- * mutex.synchronize {
- * # Thread 'b' has finished using the resource
- * resource.signal
- * }
- * }
- */
-
static size_t
condvar_memsize(const void *ptr)
{
@@ -1488,21 +1197,8 @@ condvar_alloc(VALUE klass)
return obj;
}
-/*
- * Document-method: ConditionVariable::new
- *
- * Creates a new condition variable instance.
- */
-
-static VALUE
-rb_condvar_initialize(VALUE self)
-{
- struct rb_condvar *cv = condvar_ptr(self);
- ccan_list_head_init(&cv->waitq);
- return self;
-}
-
struct sleep_call {
+ rb_execution_context_t *ec;
VALUE mutex;
VALUE timeout;
};
@@ -1513,161 +1209,72 @@ static VALUE
do_sleep(VALUE args)
{
struct sleep_call *p = (struct sleep_call *)args;
- return rb_funcallv(p->mutex, id_sleep, 1, &p->timeout);
+ if (CLASS_OF(p->mutex) == rb_cMutex) {
+ return rb_mut_sleep(p->ec, p->mutex, p->timeout);
+ }
+ else {
+ return rb_funcallv(p->mutex, id_sleep, 1, &p->timeout);
+ }
}
-/*
- * Document-method: Thread::ConditionVariable#wait
- * call-seq: wait(mutex, timeout=nil)
- *
- * Releases the lock held in +mutex+ and waits; reacquires the lock on wakeup.
- *
- * If +timeout+ is given, this method returns after +timeout+ seconds passed,
- * even if no other thread doesn't signal.
- *
- * Returns the slept result on +mutex+.
- */
-
static VALUE
-rb_condvar_wait(int argc, VALUE *argv, VALUE self)
+rb_condvar_wait(rb_execution_context_t *ec, VALUE self, VALUE mutex, VALUE timeout)
{
- rb_execution_context_t *ec = GET_EC();
-
struct rb_condvar *cv = condvar_ptr(self);
- struct sleep_call args;
-
- rb_scan_args(argc, argv, "11", &args.mutex, &args.timeout);
+ struct sleep_call args = {
+ .ec = ec,
+ .mutex = mutex,
+ .timeout = timeout,
+ };
struct sync_waiter sync_waiter = {
- .self = args.mutex,
+ .self = mutex,
.th = ec->thread_ptr,
.fiber = nonblocking_fiber(ec->fiber_ptr)
};
ccan_list_add_tail(&cv->waitq, &sync_waiter.node);
- return rb_ensure(do_sleep, (VALUE)&args, delete_from_waitq, (VALUE)&sync_waiter);
+ return rb_ec_ensure(ec, do_sleep, (VALUE)&args, delete_from_waitq, (VALUE)&sync_waiter);
}
-/*
- * Document-method: Thread::ConditionVariable#signal
- *
- * Wakes up the first thread in line waiting for this lock.
- */
-
static VALUE
-rb_condvar_signal(VALUE self)
+rb_condvar_signal(rb_execution_context_t *ec, VALUE self)
{
struct rb_condvar *cv = condvar_ptr(self);
wakeup_one(&cv->waitq);
return self;
}
-/*
- * Document-method: Thread::ConditionVariable#broadcast
- *
- * Wakes up all threads waiting for this lock.
- */
-
static VALUE
-rb_condvar_broadcast(VALUE self)
+rb_condvar_broadcast(rb_execution_context_t *ec, VALUE self)
{
struct rb_condvar *cv = condvar_ptr(self);
wakeup_all(&cv->waitq);
return self;
}
-NORETURN(static VALUE undumpable(VALUE obj));
-/* :nodoc: */
-static VALUE
-undumpable(VALUE obj)
-{
- rb_raise(rb_eTypeError, "can't dump %"PRIsVALUE, rb_obj_class(obj));
- UNREACHABLE_RETURN(Qnil);
-}
-
-static VALUE
-define_thread_class(VALUE outer, const ID name, VALUE super)
-{
- VALUE klass = rb_define_class_id_under(outer, name, super);
- rb_const_set(rb_cObject, name, klass);
- return klass;
-}
-
static void
Init_thread_sync(void)
{
-#undef rb_intern
-#if defined(TEACH_RDOC) && TEACH_RDOC == 42
- rb_cMutex = rb_define_class_under(rb_cThread, "Mutex", rb_cObject);
- rb_cConditionVariable = rb_define_class_under(rb_cThread, "ConditionVariable", rb_cObject);
- rb_cQueue = rb_define_class_under(rb_cThread, "Queue", rb_cObject);
- rb_cSizedQueue = rb_define_class_under(rb_cThread, "SizedQueue", rb_cObject);
-#endif
-
-#define DEFINE_CLASS(name, super) \
- rb_c##name = define_thread_class(rb_cThread, rb_intern(#name), rb_c##super)
-
/* Mutex */
- DEFINE_CLASS(Mutex, Object);
+ rb_cMutex = rb_define_class_id_under(rb_cThread, rb_intern("Mutex"), rb_cObject);
rb_define_alloc_func(rb_cMutex, mutex_alloc);
- rb_define_method(rb_cMutex, "initialize", mutex_initialize, 0);
- rb_define_method(rb_cMutex, "locked?", rb_mutex_locked_p, 0);
- rb_define_method(rb_cMutex, "try_lock", rb_mutex_trylock, 0);
- rb_define_method(rb_cMutex, "lock", rb_mutex_lock, 0);
- rb_define_method(rb_cMutex, "unlock", rb_mutex_unlock, 0);
- rb_define_method(rb_cMutex, "sleep", mutex_sleep, -1);
- rb_define_method(rb_cMutex, "synchronize", rb_mutex_synchronize_m, 0);
- rb_define_method(rb_cMutex, "owned?", rb_mutex_owned_p, 0);
/* Queue */
- DEFINE_CLASS(Queue, Object);
+ VALUE rb_cQueue = rb_define_class_id_under_no_pin(rb_cThread, rb_intern("Queue"), rb_cObject);
rb_define_alloc_func(rb_cQueue, queue_alloc);
rb_eClosedQueueError = rb_define_class("ClosedQueueError", rb_eStopIteration);
- rb_define_method(rb_cQueue, "initialize", rb_queue_initialize, -1);
- rb_undef_method(rb_cQueue, "initialize_copy");
- rb_define_method(rb_cQueue, "marshal_dump", undumpable, 0);
- rb_define_method(rb_cQueue, "close", rb_queue_close, 0);
- rb_define_method(rb_cQueue, "closed?", rb_queue_closed_p, 0);
- rb_define_method(rb_cQueue, "push", rb_queue_push, 1);
- rb_define_method(rb_cQueue, "empty?", rb_queue_empty_p, 0);
- rb_define_method(rb_cQueue, "clear", rb_queue_clear, 0);
- rb_define_method(rb_cQueue, "length", rb_queue_length, 0);
- rb_define_method(rb_cQueue, "num_waiting", rb_queue_num_waiting, 0);
- rb_define_method(rb_cQueue, "freeze", rb_queue_freeze, 0);
-
- rb_define_alias(rb_cQueue, "enq", "push");
- rb_define_alias(rb_cQueue, "<<", "push");
- rb_define_alias(rb_cQueue, "size", "length");
-
- DEFINE_CLASS(SizedQueue, Queue);
+ VALUE rb_cSizedQueue = rb_define_class_id_under_no_pin(rb_cThread, rb_intern("SizedQueue"), rb_cQueue);
rb_define_alloc_func(rb_cSizedQueue, szqueue_alloc);
- rb_define_method(rb_cSizedQueue, "initialize", rb_szqueue_initialize, 1);
- rb_define_method(rb_cSizedQueue, "close", rb_szqueue_close, 0);
- rb_define_method(rb_cSizedQueue, "max", rb_szqueue_max_get, 0);
- rb_define_method(rb_cSizedQueue, "max=", rb_szqueue_max_set, 1);
- rb_define_method(rb_cSizedQueue, "empty?", rb_szqueue_empty_p, 0);
- rb_define_method(rb_cSizedQueue, "clear", rb_szqueue_clear, 0);
- rb_define_method(rb_cSizedQueue, "length", rb_szqueue_length, 0);
- rb_define_method(rb_cSizedQueue, "num_waiting", rb_szqueue_num_waiting, 0);
- rb_define_method(rb_cSizedQueue, "freeze", rb_queue_freeze, 0);
- rb_define_alias(rb_cSizedQueue, "size", "length");
-
/* CVar */
- DEFINE_CLASS(ConditionVariable, Object);
+ VALUE rb_cConditionVariable = rb_define_class_id_under_no_pin(rb_cThread, rb_intern("ConditionVariable"), rb_cObject);
rb_define_alloc_func(rb_cConditionVariable, condvar_alloc);
id_sleep = rb_intern("sleep");
- rb_define_method(rb_cConditionVariable, "initialize", rb_condvar_initialize, 0);
- rb_undef_method(rb_cConditionVariable, "initialize_copy");
- rb_define_method(rb_cConditionVariable, "marshal_dump", undumpable, 0);
- rb_define_method(rb_cConditionVariable, "wait", rb_condvar_wait, -1);
- rb_define_method(rb_cConditionVariable, "signal", rb_condvar_signal, 0);
- rb_define_method(rb_cConditionVariable, "broadcast", rb_condvar_broadcast, 0);
-
rb_provide("thread.rb");
}