mirror of
https://github.com/ruby/ruby.git
synced 2022-11-09 12:17:21 -05:00
* thread_pthread.c (register_ubf_list): renamed from
add_signal_thread_list. * thread_pthread.c (unregister_ubf_list): renamed from remove_signal_thread_list. * thread_pthread.c (ubf_wakeup_all_threads): renamed from ping_signal_thread_list. * thread_pthread.c (ubf_wakeup_thread): renamed from ubf_select_each. * thread_pthread.c (ubf_threads_empty): renamed from check_signal_thread_list(). * thread_pthread.c (ubf_list_lock): renamed from signal_thread_list_lock. * thread_pthread.c (register_ubf_list): large simplification by using ccan/list.h. bonus: removed malloc() and exit(EXIT_FAILURE). * thread_pthread.c (unregister_ubf_list): ditto. * thread_pthread.c (ubf_threads_empty): ditto. * thread_pthread.c (ubf_wakeup_all_threads): ditto. * thread_pthread.c (print_signal_list): removed. git-svn-id: svn+ssh://ci.ruby-lang.org/ruby/trunk@52818 b2dd03c8-39d4-4d8f-98ff-823fe69b080e
This commit is contained in:
parent
a014e9e8c6
commit
487748fac8
5 changed files with 91 additions and 117 deletions
24
ChangeLog
24
ChangeLog
|
@ -1,3 +1,27 @@
|
||||||
|
Thu Oct 22 12:54:43 2015 KOSAKI Motohiro <kosaki.motohiro@gmail.com>
|
||||||
|
|
||||||
|
* thread_pthread.c (register_ubf_list): renamed from
|
||||||
|
add_signal_thread_list.
|
||||||
|
* thread_pthread.c (unregister_ubf_list): renamed
|
||||||
|
from remove_signal_thread_list.
|
||||||
|
* thread_pthread.c (ubf_wakeup_all_threads): renamed from
|
||||||
|
ping_signal_thread_list.
|
||||||
|
* thread_pthread.c (ubf_wakeup_thread): renamed from
|
||||||
|
ubf_select_each.
|
||||||
|
* thread_pthread.c (ubf_threads_empty): renamed from
|
||||||
|
check_signal_thread_list().
|
||||||
|
* thread_pthread.c (ubf_list_lock): renamed from
|
||||||
|
signal_thread_list_lock.
|
||||||
|
|
||||||
|
* thread_pthread.c (register_ubf_list): large simplification
|
||||||
|
by using ccan/list.h.
|
||||||
|
bonus: removed malloc() and exit(EXIT_FAILURE).
|
||||||
|
* thread_pthread.c (unregister_ubf_list): ditto.
|
||||||
|
* thread_pthread.c (ubf_threads_empty): ditto.
|
||||||
|
* thread_pthread.c (ubf_wakeup_all_threads): ditto.
|
||||||
|
|
||||||
|
* thread_pthread.c (print_signal_list): removed.
|
||||||
|
|
||||||
Thu Oct 22 08:03:49 2015 KOSAKI Motohiro <kosaki.motohiro@gmail.com>
|
Thu Oct 22 08:03:49 2015 KOSAKI Motohiro <kosaki.motohiro@gmail.com>
|
||||||
|
|
||||||
* test/ruby/test_rand.rb (TestRand#test_default_seed): add
|
* test/ruby/test_rand.rb (TestRand#test_default_seed): add
|
||||||
|
|
2
thread.c
2
thread.c
|
@ -1261,7 +1261,7 @@ blocking_region_end(rb_thread_t *th, struct rb_blocking_region_buffer *region)
|
||||||
gvl_acquire(th->vm, th);
|
gvl_acquire(th->vm, th);
|
||||||
rb_thread_set_current(th);
|
rb_thread_set_current(th);
|
||||||
thread_debug("leave blocking region (%p)\n", (void *)th);
|
thread_debug("leave blocking region (%p)\n", (void *)th);
|
||||||
remove_signal_thread_list(th);
|
unregister_ubf_list(th);
|
||||||
th->blocking_region_buffer = 0;
|
th->blocking_region_buffer = 0;
|
||||||
reset_unblock_function(th, ®ion->oldubf);
|
reset_unblock_function(th, ®ion->oldubf);
|
||||||
if (th->status == THREAD_STOPPED) {
|
if (th->status == THREAD_STOPPED) {
|
||||||
|
|
178
thread_pthread.c
178
thread_pthread.c
|
@ -420,12 +420,8 @@ native_cond_timeout(rb_nativethread_cond_t *cond, struct timespec timeout_rel)
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#if defined(SIGVTALRM) && !defined(__CYGWIN__)
|
#if defined(SIGVTALRM) && !defined(__CYGWIN__)
|
||||||
#define USE_SIGNAL_THREAD_LIST 1
|
#define USE_UBF_LIST 1
|
||||||
#endif
|
static rb_nativethread_lock_t ubf_list_lock;
|
||||||
#ifdef USE_SIGNAL_THREAD_LIST
|
|
||||||
static void add_signal_thread_list(rb_thread_t *th);
|
|
||||||
static void remove_signal_thread_list(rb_thread_t *th);
|
|
||||||
static rb_nativethread_lock_t signal_thread_list_lock;
|
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
static pthread_key_t ruby_native_thread_key;
|
static pthread_key_t ruby_native_thread_key;
|
||||||
|
@ -459,8 +455,8 @@ Init_native_thread(void)
|
||||||
th->thread_id = pthread_self();
|
th->thread_id = pthread_self();
|
||||||
fill_thread_id_str(th);
|
fill_thread_id_str(th);
|
||||||
native_thread_init(th);
|
native_thread_init(th);
|
||||||
#ifdef USE_SIGNAL_THREAD_LIST
|
#ifdef USE_UBF_LIST
|
||||||
native_mutex_initialize(&signal_thread_list_lock);
|
native_mutex_initialize(&ubf_list_lock);
|
||||||
#endif
|
#endif
|
||||||
#ifndef __native_client__
|
#ifndef __native_client__
|
||||||
posix_signal(SIGVTALRM, null_func);
|
posix_signal(SIGVTALRM, null_func);
|
||||||
|
@ -470,7 +466,12 @@ Init_native_thread(void)
|
||||||
static void
|
static void
|
||||||
native_thread_init(rb_thread_t *th)
|
native_thread_init(rb_thread_t *th)
|
||||||
{
|
{
|
||||||
native_cond_initialize(&th->native_thread_data.sleep_cond, RB_CONDATTR_CLOCK_MONOTONIC);
|
native_thread_data_t *nd = &th->native_thread_data;
|
||||||
|
|
||||||
|
#ifdef USE_UBF_LIST
|
||||||
|
list_node_init(&nd->ubf_list);
|
||||||
|
#endif
|
||||||
|
native_cond_initialize(&nd->sleep_cond, RB_CONDATTR_CLOCK_MONOTONIC);
|
||||||
ruby_thread_set_native(th);
|
ruby_thread_set_native(th);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1135,144 +1136,93 @@ native_sleep(rb_thread_t *th, struct timeval *timeout_tv)
|
||||||
thread_debug("native_sleep done\n");
|
thread_debug("native_sleep done\n");
|
||||||
}
|
}
|
||||||
|
|
||||||
#ifdef USE_SIGNAL_THREAD_LIST
|
#ifdef USE_UBF_LIST
|
||||||
struct signal_thread_list {
|
static LIST_HEAD(ubf_list_head);
|
||||||
rb_thread_t *th;
|
|
||||||
struct signal_thread_list *prev;
|
|
||||||
struct signal_thread_list *next;
|
|
||||||
};
|
|
||||||
|
|
||||||
static struct signal_thread_list signal_thread_list_anchor = {
|
/* The thread 'th' is registered to be trying unblock. */
|
||||||
0, 0, 0,
|
|
||||||
};
|
|
||||||
|
|
||||||
#define FGLOCK(lock, body) do { \
|
|
||||||
native_mutex_lock(lock); \
|
|
||||||
{ \
|
|
||||||
body; \
|
|
||||||
} \
|
|
||||||
native_mutex_unlock(lock); \
|
|
||||||
} while (0)
|
|
||||||
|
|
||||||
#if 0 /* for debug */
|
|
||||||
static void
|
static void
|
||||||
print_signal_list(char *str)
|
register_ubf_list(rb_thread_t *th)
|
||||||
{
|
{
|
||||||
struct signal_thread_list *list =
|
struct list_node *node = &th->native_thread_data.ubf_list;
|
||||||
signal_thread_list_anchor.next;
|
|
||||||
thread_debug("list (%s)> ", str);
|
|
||||||
while (list) {
|
|
||||||
thread_debug("%p (%p), ", list->th, list->th->thread_id);
|
|
||||||
list = list->next;
|
|
||||||
}
|
|
||||||
thread_debug("\n");
|
|
||||||
}
|
|
||||||
#endif
|
|
||||||
|
|
||||||
static void
|
if (list_empty((struct list_head*)node)) {
|
||||||
add_signal_thread_list(rb_thread_t *th)
|
native_mutex_lock(&ubf_list_lock);
|
||||||
{
|
list_add(&ubf_list_head, node);
|
||||||
if (!th->native_thread_data.signal_thread_list) {
|
native_mutex_unlock(&ubf_list_lock);
|
||||||
FGLOCK(&signal_thread_list_lock, {
|
|
||||||
struct signal_thread_list *list =
|
|
||||||
malloc(sizeof(struct signal_thread_list));
|
|
||||||
|
|
||||||
if (list == 0) {
|
|
||||||
fprintf(stderr, "[FATAL] failed to allocate memory\n");
|
|
||||||
exit(EXIT_FAILURE);
|
|
||||||
}
|
|
||||||
|
|
||||||
list->th = th;
|
|
||||||
|
|
||||||
list->prev = &signal_thread_list_anchor;
|
|
||||||
list->next = signal_thread_list_anchor.next;
|
|
||||||
if (list->next) {
|
|
||||||
list->next->prev = list;
|
|
||||||
}
|
|
||||||
signal_thread_list_anchor.next = list;
|
|
||||||
th->native_thread_data.signal_thread_list = list;
|
|
||||||
});
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* The thread 'th' is unblocked. It no longer need to be registered. */
|
||||||
static void
|
static void
|
||||||
remove_signal_thread_list(rb_thread_t *th)
|
unregister_ubf_list(rb_thread_t *th)
|
||||||
{
|
{
|
||||||
if (th->native_thread_data.signal_thread_list) {
|
struct list_node *node = &th->native_thread_data.ubf_list;
|
||||||
FGLOCK(&signal_thread_list_lock, {
|
|
||||||
struct signal_thread_list *list =
|
|
||||||
(struct signal_thread_list *)
|
|
||||||
th->native_thread_data.signal_thread_list;
|
|
||||||
|
|
||||||
list->prev->next = list->next;
|
if (!list_empty((struct list_head*)node)) {
|
||||||
if (list->next) {
|
native_mutex_lock(&ubf_list_lock);
|
||||||
list->next->prev = list->prev;
|
list_del_init(node);
|
||||||
}
|
native_mutex_unlock(&ubf_list_lock);
|
||||||
th->native_thread_data.signal_thread_list = 0;
|
|
||||||
list->th = 0;
|
|
||||||
free(list); /* ok */
|
|
||||||
});
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* send a signal to intent that a target thread return from blocking syscall.
|
||||||
|
* Maybe any signal is ok, but we chose SIGVTALRM.
|
||||||
|
*/
|
||||||
static void
|
static void
|
||||||
ubf_select_each(rb_thread_t *th)
|
ubf_wakeup_thread(rb_thread_t *th)
|
||||||
{
|
{
|
||||||
thread_debug("ubf_select_each (%"PRI_THREAD_ID")\n", thread_id_str(th));
|
thread_debug("thread_wait_queue_wakeup (%"PRI_THREAD_ID")\n", thread_id_str(th));
|
||||||
if (th) {
|
if (th)
|
||||||
pthread_kill(th->thread_id, SIGVTALRM);
|
pthread_kill(th->thread_id, SIGVTALRM);
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void
|
static void
|
||||||
ubf_select(void *ptr)
|
ubf_select(void *ptr)
|
||||||
{
|
{
|
||||||
rb_thread_t *th = (rb_thread_t *)ptr;
|
rb_thread_t *th = (rb_thread_t *)ptr;
|
||||||
add_signal_thread_list(th);
|
register_ubf_list(th);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* ubf_select_each() doesn't guarantee to wake up the target thread.
|
* ubf_wakeup_thread() doesn't guarantee to wake up a target thread.
|
||||||
* Therefore, we need to activate timer thread when called from
|
* Therefore, we repeatedly call ubf_wakeup_thread() until a target thread
|
||||||
* Thread#kill etc.
|
* exit from ubf function.
|
||||||
* In the other hands, we shouldn't call rb_thread_wakeup_timer_thread()
|
* In the other hands, we shouldn't call rb_thread_wakeup_timer_thread()
|
||||||
* if running on timer thread because it may make endless wakeups.
|
* if running on timer thread because it may make endless wakeups.
|
||||||
*/
|
*/
|
||||||
if (!pthread_equal(pthread_self(), timer_thread.id))
|
if (!pthread_equal(pthread_self(), timer_thread.id))
|
||||||
rb_thread_wakeup_timer_thread();
|
rb_thread_wakeup_timer_thread();
|
||||||
ubf_select_each(th);
|
ubf_wakeup_thread(th);
|
||||||
}
|
|
||||||
|
|
||||||
static void
|
|
||||||
ping_signal_thread_list(void)
|
|
||||||
{
|
|
||||||
if (signal_thread_list_anchor.next) {
|
|
||||||
FGLOCK(&signal_thread_list_lock, {
|
|
||||||
struct signal_thread_list *list;
|
|
||||||
|
|
||||||
list = signal_thread_list_anchor.next;
|
|
||||||
while (list) {
|
|
||||||
ubf_select_each(list->th);
|
|
||||||
list = list->next;
|
|
||||||
}
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static int
|
static int
|
||||||
check_signal_thread_list(void)
|
ubf_threads_empty(void)
|
||||||
{
|
{
|
||||||
if (signal_thread_list_anchor.next)
|
return list_empty(&ubf_list_head);
|
||||||
return 1;
|
|
||||||
else
|
|
||||||
return 0;
|
|
||||||
}
|
}
|
||||||
#else /* USE_SIGNAL_THREAD_LIST */
|
|
||||||
#define add_signal_thread_list(th) (void)(th)
|
static void
|
||||||
#define remove_signal_thread_list(th) (void)(th)
|
ubf_wakeup_all_threads(void)
|
||||||
|
{
|
||||||
|
rb_thread_t *th;
|
||||||
|
|
||||||
|
if (!ubf_threads_empty()) {
|
||||||
|
native_mutex_lock(&ubf_list_lock);
|
||||||
|
list_for_each(&ubf_list_head, th,
|
||||||
|
native_thread_data.ubf_list) {
|
||||||
|
ubf_wakeup_thread(th);
|
||||||
|
}
|
||||||
|
native_mutex_unlock(&ubf_list_lock);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#else /* USE_UBF_LIST */
|
||||||
|
#define register_ubf_list(th) (void)(th)
|
||||||
|
#define unregister_ubf_list(th) (void)(th)
|
||||||
#define ubf_select 0
|
#define ubf_select 0
|
||||||
static void ping_signal_thread_list(void) { return; }
|
static void ubf_wakeup_all_threads(void) { return; }
|
||||||
static int check_signal_thread_list(void) { return 0; }
|
static int ubf_threads_empty(void) { return 1; }
|
||||||
#endif /* USE_SIGNAL_THREAD_LIST */
|
#endif /* USE_UBF_LIST */
|
||||||
|
|
||||||
#define TT_DEBUG 0
|
#define TT_DEBUG 0
|
||||||
#define WRITE_CONST(fd, str) (void)(write((fd),(str),sizeof(str)-1)<0)
|
#define WRITE_CONST(fd, str) (void)(write((fd),(str),sizeof(str)-1)<0)
|
||||||
|
@ -1479,7 +1429,7 @@ timer_thread_sleep(rb_global_vm_lock_t* gvl)
|
||||||
pollfds[1].fd = timer_thread_pipe.low[0];
|
pollfds[1].fd = timer_thread_pipe.low[0];
|
||||||
pollfds[1].events = POLLIN;
|
pollfds[1].events = POLLIN;
|
||||||
|
|
||||||
need_polling = check_signal_thread_list();
|
need_polling = !ubf_threads_empty();
|
||||||
|
|
||||||
if (gvl->waiting > 0 || need_polling) {
|
if (gvl->waiting > 0 || need_polling) {
|
||||||
/* polling (TIME_QUANTUM_USEC usec) */
|
/* polling (TIME_QUANTUM_USEC usec) */
|
||||||
|
@ -1587,7 +1537,7 @@ thread_timer(void *p)
|
||||||
while (system_working > 0) {
|
while (system_working > 0) {
|
||||||
|
|
||||||
/* timer function */
|
/* timer function */
|
||||||
ping_signal_thread_list();
|
ubf_wakeup_all_threads();
|
||||||
timer_thread_function(0);
|
timer_thread_function(0);
|
||||||
|
|
||||||
if (TT_DEBUG) WRITE_CONST(2, "tick\n");
|
if (TT_DEBUG) WRITE_CONST(2, "tick\n");
|
||||||
|
|
|
@ -26,7 +26,7 @@ typedef struct rb_thread_cond_struct {
|
||||||
} rb_nativethread_cond_t;
|
} rb_nativethread_cond_t;
|
||||||
|
|
||||||
typedef struct native_thread_data_struct {
|
typedef struct native_thread_data_struct {
|
||||||
void *signal_thread_list;
|
struct list_node ubf_list;
|
||||||
rb_nativethread_cond_t sleep_cond;
|
rb_nativethread_cond_t sleep_cond;
|
||||||
} native_thread_data_t;
|
} native_thread_data_t;
|
||||||
|
|
||||||
|
|
|
@ -19,7 +19,7 @@
|
||||||
#undef Sleep
|
#undef Sleep
|
||||||
|
|
||||||
#define native_thread_yield() Sleep(0)
|
#define native_thread_yield() Sleep(0)
|
||||||
#define remove_signal_thread_list(th)
|
#define unregister_ubf_list(th)
|
||||||
|
|
||||||
static volatile DWORD ruby_native_thread_key = TLS_OUT_OF_INDEXES;
|
static volatile DWORD ruby_native_thread_key = TLS_OUT_OF_INDEXES;
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue