summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--ChangeLog24
-rw-r--r--thread.c2
-rw-r--r--thread_pthread.c172
-rw-r--r--thread_pthread.h2
-rw-r--r--thread_win32.c2
5 files changed, 88 insertions, 114 deletions
diff --git a/ChangeLog b/ChangeLog
index 30c0f9b..5b7adcb 100644
--- a/ChangeLog
+++ b/ChangeLog
@@ -1,3 +1,27 @@
+Thu Oct 22 12:54:43 2015 KOSAKI Motohiro <kosaki.motohiro@gmail.com>
+
+ * thread_pthread.c (register_ubf_list): renamed from
+ add_signal_thread_list.
+ * thread_pthread.c (unregister_ubf_list): renamed
+ from remove_signal_thread_list.
+ * thread_pthread.c (ubf_wakeup_all_threads): renamed from
+ ping_signal_thread_list.
+ * thread_pthread.c (ubf_wakeup_thread): renamed from
+ ubf_select_each.
+ * thread_pthread.c (ubf_threads_empty): renamed from
+ check_signal_thread_list().
+ * thread_pthread.c (ubf_list_lock): renamed from
+ signal_thread_list_lock.
+
+ * thread_pthread.c (register_ubf_list): large simplification
+ by using ccan/list.h.
+ bonus: removed malloc() and exit(EXIT_FAILURE).
+ * thread_pthread.c (unregister_ubf_list): ditto.
+ * thread_pthread.c (ubf_threads_empty): ditto.
+ * thread_pthread.c (ubf_wakeup_all_threads): ditto.
+
+ * thread_pthread.c (print_signal_list): removed.
+
Thu Oct 22 08:03:49 2015 KOSAKI Motohiro <kosaki.motohiro@gmail.com>
* test/ruby/test_rand.rb (TestRand#test_default_seed): add
diff --git a/thread.c b/thread.c
index bed6551..7190a94 100644
--- a/thread.c
+++ b/thread.c
@@ -1261,7 +1261,7 @@ blocking_region_end(rb_thread_t *th, struct rb_blocking_region_buffer *region)
gvl_acquire(th->vm, th);
rb_thread_set_current(th);
thread_debug("leave blocking region (%p)\n", (void *)th);
- remove_signal_thread_list(th);
+ unregister_ubf_list(th);
th->blocking_region_buffer = 0;
reset_unblock_function(th, &region->oldubf);
if (th->status == THREAD_STOPPED) {
diff --git a/thread_pthread.c b/thread_pthread.c
index 39271c8..3804590 100644
--- a/thread_pthread.c
+++ b/thread_pthread.c
@@ -420,12 +420,8 @@ native_cond_timeout(rb_nativethread_cond_t *cond, struct timespec timeout_rel)
#endif
#if defined(SIGVTALRM) && !defined(__CYGWIN__)
-#define USE_SIGNAL_THREAD_LIST 1
-#endif
-#ifdef USE_SIGNAL_THREAD_LIST
-static void add_signal_thread_list(rb_thread_t *th);
-static void remove_signal_thread_list(rb_thread_t *th);
-static rb_nativethread_lock_t signal_thread_list_lock;
+#define USE_UBF_LIST 1
+static rb_nativethread_lock_t ubf_list_lock;
#endif
static pthread_key_t ruby_native_thread_key;
@@ -459,8 +455,8 @@ Init_native_thread(void)
th->thread_id = pthread_self();
fill_thread_id_str(th);
native_thread_init(th);
-#ifdef USE_SIGNAL_THREAD_LIST
- native_mutex_initialize(&signal_thread_list_lock);
+#ifdef USE_UBF_LIST
+ native_mutex_initialize(&ubf_list_lock);
#endif
#ifndef __native_client__
posix_signal(SIGVTALRM, null_func);
@@ -470,7 +466,12 @@ Init_native_thread(void)
static void
native_thread_init(rb_thread_t *th)
{
- native_cond_initialize(&th->native_thread_data.sleep_cond, RB_CONDATTR_CLOCK_MONOTONIC);
+ native_thread_data_t *nd = &th->native_thread_data;
+
+#ifdef USE_UBF_LIST
+ list_node_init(&nd->ubf_list);
+#endif
+ native_cond_initialize(&nd->sleep_cond, RB_CONDATTR_CLOCK_MONOTONIC);
ruby_thread_set_native(th);
}
@@ -1135,144 +1136,93 @@ native_sleep(rb_thread_t *th, struct timeval *timeout_tv)
thread_debug("native_sleep done\n");
}
-#ifdef USE_SIGNAL_THREAD_LIST
-struct signal_thread_list {
- rb_thread_t *th;
- struct signal_thread_list *prev;
- struct signal_thread_list *next;
-};
-
-static struct signal_thread_list signal_thread_list_anchor = {
- 0, 0, 0,
-};
-
-#define FGLOCK(lock, body) do { \
- native_mutex_lock(lock); \
- { \
- body; \
- } \
- native_mutex_unlock(lock); \
-} while (0)
-
-#if 0 /* for debug */
-static void
-print_signal_list(char *str)
-{
- struct signal_thread_list *list =
- signal_thread_list_anchor.next;
- thread_debug("list (%s)> ", str);
- while (list) {
- thread_debug("%p (%p), ", list->th, list->th->thread_id);
- list = list->next;
- }
- thread_debug("\n");
-}
-#endif
+#ifdef USE_UBF_LIST
+static LIST_HEAD(ubf_list_head);
+/* The thread 'th' is registered to be trying unblock. */
static void
-add_signal_thread_list(rb_thread_t *th)
+register_ubf_list(rb_thread_t *th)
{
- if (!th->native_thread_data.signal_thread_list) {
- FGLOCK(&signal_thread_list_lock, {
- struct signal_thread_list *list =
- malloc(sizeof(struct signal_thread_list));
+ struct list_node *node = &th->native_thread_data.ubf_list;
- if (list == 0) {
- fprintf(stderr, "[FATAL] failed to allocate memory\n");
- exit(EXIT_FAILURE);
- }
-
- list->th = th;
-
- list->prev = &signal_thread_list_anchor;
- list->next = signal_thread_list_anchor.next;
- if (list->next) {
- list->next->prev = list;
- }
- signal_thread_list_anchor.next = list;
- th->native_thread_data.signal_thread_list = list;
- });
+ if (list_empty((struct list_head*)node)) {
+ native_mutex_lock(&ubf_list_lock);
+ list_add(&ubf_list_head, node);
+ native_mutex_unlock(&ubf_list_lock);
}
}
+/* The thread 'th' is unblocked. It no longer need to be registered. */
static void
-remove_signal_thread_list(rb_thread_t *th)
+unregister_ubf_list(rb_thread_t *th)
{
- if (th->native_thread_data.signal_thread_list) {
- FGLOCK(&signal_thread_list_lock, {
- struct signal_thread_list *list =
- (struct signal_thread_list *)
- th->native_thread_data.signal_thread_list;
+ struct list_node *node = &th->native_thread_data.ubf_list;
- list->prev->next = list->next;
- if (list->next) {
- list->next->prev = list->prev;
- }
- th->native_thread_data.signal_thread_list = 0;
- list->th = 0;
- free(list); /* ok */
- });
+ if (!list_empty((struct list_head*)node)) {
+ native_mutex_lock(&ubf_list_lock);
+ list_del_init(node);
+ native_mutex_unlock(&ubf_list_lock);
}
}
+/*
+ * send a signal to intent that a target thread return from blocking syscall.
+ * Maybe any signal is ok, but we chose SIGVTALRM.
+ */
static void
-ubf_select_each(rb_thread_t *th)
+ubf_wakeup_thread(rb_thread_t *th)
{
- thread_debug("ubf_select_each (%"PRI_THREAD_ID")\n", thread_id_str(th));
- if (th) {
+ thread_debug("thread_wait_queue_wakeup (%"PRI_THREAD_ID")\n", thread_id_str(th));
+ if (th)
pthread_kill(th->thread_id, SIGVTALRM);
- }
}
static void
ubf_select(void *ptr)
{
rb_thread_t *th = (rb_thread_t *)ptr;
- add_signal_thread_list(th);
+ register_ubf_list(th);
/*
- * ubf_select_each() doesn't guarantee to wake up the target thread.
- * Therefore, we need to activate timer thread when called from
- * Thread#kill etc.
+ * ubf_wakeup_thread() doesn't guarantee to wake up a target thread.
+ * Therefore, we repeatedly call ubf_wakeup_thread() until a target thread
+ * exit from ubf function.
* In the other hands, we shouldn't call rb_thread_wakeup_timer_thread()
* if running on timer thread because it may make endless wakeups.
*/
if (!pthread_equal(pthread_self(), timer_thread.id))
rb_thread_wakeup_timer_thread();
- ubf_select_each(th);
+ ubf_wakeup_thread(th);
+}
+
+static int
+ubf_threads_empty(void)
+{
+ return list_empty(&ubf_list_head);
}
static void
-ping_signal_thread_list(void)
+ubf_wakeup_all_threads(void)
{
- if (signal_thread_list_anchor.next) {
- FGLOCK(&signal_thread_list_lock, {
- struct signal_thread_list *list;
+ rb_thread_t *th;
- list = signal_thread_list_anchor.next;
- while (list) {
- ubf_select_each(list->th);
- list = list->next;
- }
- });
+ if (!ubf_threads_empty()) {
+ native_mutex_lock(&ubf_list_lock);
+ list_for_each(&ubf_list_head, th,
+ native_thread_data.ubf_list) {
+ ubf_wakeup_thread(th);
+ }
+ native_mutex_unlock(&ubf_list_lock);
}
}
-static int
-check_signal_thread_list(void)
-{
- if (signal_thread_list_anchor.next)
- return 1;
- else
- return 0;
-}
-#else /* USE_SIGNAL_THREAD_LIST */
-#define add_signal_thread_list(th) (void)(th)
-#define remove_signal_thread_list(th) (void)(th)
+#else /* USE_UBF_LIST */
+#define register_ubf_list(th) (void)(th)
+#define unregister_ubf_list(th) (void)(th)
#define ubf_select 0
-static void ping_signal_thread_list(void) { return; }
-static int check_signal_thread_list(void) { return 0; }
-#endif /* USE_SIGNAL_THREAD_LIST */
+static void ubf_wakeup_all_threads(void) { return; }
+static int ubf_threads_empty(void) { return 1; }
+#endif /* USE_UBF_LIST */
#define TT_DEBUG 0
#define WRITE_CONST(fd, str) (void)(write((fd),(str),sizeof(str)-1)<0)
@@ -1479,7 +1429,7 @@ timer_thread_sleep(rb_global_vm_lock_t* gvl)
pollfds[1].fd = timer_thread_pipe.low[0];
pollfds[1].events = POLLIN;
- need_polling = check_signal_thread_list();
+ need_polling = !ubf_threads_empty();
if (gvl->waiting > 0 || need_polling) {
/* polling (TIME_QUANTUM_USEC usec) */
@@ -1587,7 +1537,7 @@ thread_timer(void *p)
while (system_working > 0) {
/* timer function */
- ping_signal_thread_list();
+ ubf_wakeup_all_threads();
timer_thread_function(0);
if (TT_DEBUG) WRITE_CONST(2, "tick\n");
diff --git a/thread_pthread.h b/thread_pthread.h
index 24a4af4..1c3782c 100644
--- a/thread_pthread.h
+++ b/thread_pthread.h
@@ -26,7 +26,7 @@ typedef struct rb_thread_cond_struct {
} rb_nativethread_cond_t;
typedef struct native_thread_data_struct {
- void *signal_thread_list;
+ struct list_node ubf_list;
rb_nativethread_cond_t sleep_cond;
} native_thread_data_t;
diff --git a/thread_win32.c b/thread_win32.c
index 2dfbf06..d9a9bd1 100644
--- a/thread_win32.c
+++ b/thread_win32.c
@@ -19,7 +19,7 @@
#undef Sleep
#define native_thread_yield() Sleep(0)
-#define remove_signal_thread_list(th)
+#define unregister_ubf_list(th)
static volatile DWORD ruby_native_thread_key = TLS_OUT_OF_INDEXES;