1
0
Fork 0
mirror of https://github.com/ruby/ruby.git synced 2022-11-09 12:17:21 -05:00

* thread.c, vm_core.h, vm.c, thread_pthread.c, thread_win32.c: add

deadlock detection.  [ruby-dev:35044]

* bootstraptest/test_thread.rb: add tests for above.


git-svn-id: svn+ssh://ci.ruby-lang.org/ruby/trunk@17110 b2dd03c8-39d4-4d8f-98ff-823fe69b080e
This commit is contained in:
mame 2008-06-12 13:01:38 +00:00
parent 2b66844f48
commit 6f5aaff73b
7 changed files with 294 additions and 40 deletions

View file

@ -1,3 +1,10 @@
Thu Jun 12 21:59:17 2008 Yusuke Endoh <mame@tsg.ne.jp>
* thread.c, vm_core.h, vm.c, thread_pthread.c, thread_win32.c: add
deadlock detection. [ruby-dev:35044]
* bootstraptest/test_thread.rb: add tests for above.
Thu Jun 12 21:39:55 2008 Tadayoshi Funaba <tadf@dotrb.org>
* complex.c: refactoring.

View file

@ -302,3 +302,56 @@ assert_normal_exit %q{
assert_normal_exit %q{
Thread.new("foo", &Object.method(:class_eval)).join
}, '[ruby-dev:34128]'
assert_equal 'ok', %q{
begin
Thread.new { Thread.stop }
Thread.stop
:ng
rescue Exception
:ok
end
}
assert_equal 'ok', %q{
begin
m1, m2 = Mutex.new, Mutex.new
Thread.new { m1.lock; sleep 1; m2.lock }
m2.lock; sleep 1; m1.lock
:ng
rescue Exception
:ok
end
}
assert_equal 'ok', %q{
m = Mutex.new
Thread.new { m.lock }; sleep 1; m.lock
:ok
}
assert_equal 'ok', %q{
m = Mutex.new
Thread.new { m.lock }; m.lock
:ok
}
assert_equal 'ok', %q{
m = Mutex.new
Thread.new { m.lock }.join; m.lock
:ok
}
assert_equal 'ok', %q{
m = Mutex.new
Thread.new { m.lock; sleep 2 }
sleep 1; m.lock
:ok
}
assert_equal 'ok', %q{
m = Mutex.new
Thread.new { m.lock; sleep 2; m.unlock }
sleep 1; m.lock
:ok
}

233
thread.c
View file

@ -57,11 +57,14 @@ VALUE rb_cBarrier;
static void sleep_timeval(rb_thread_t *th, struct timeval time);
static void sleep_wait_for_interrupt(rb_thread_t *th, double sleepsec);
static void sleep_forever(rb_thread_t *th);
static void sleep_forever(rb_thread_t *th, int nodeadlock);
static double timeofday(void);
struct timeval rb_time_interval(VALUE);
static int rb_thread_dead(rb_thread_t *th);
static void rb_mutex_unlock_all(VALUE);
static void rb_check_deadlock(rb_vm_t *vm);
void rb_signal_exec(rb_thread_t *th, int sig);
void rb_disable_interrupt(void);
@ -93,12 +96,12 @@ static void reset_unblock_function(rb_thread_t *th, const struct rb_unblock_call
rb_thread_set_current(_th_stored); \
} while(0)
#define BLOCKING_REGION(exec, ubf, ubfarg) do { \
#define BLOCKING_REGION(exec, ubf, ubfarg, stopped) do { \
rb_thread_t *__th = GET_THREAD(); \
int __prev_status = __th->status; \
struct rb_unblock_callback __oldubf; \
set_unblock_function(__th, ubf, ubfarg, &__oldubf); \
__th->status = THREAD_STOPPED; \
if (stopped) __th->status = THREAD_STOPPED; \
thread_debug("enter blocking region (%p)\n", __th); \
GVL_UNLOCK_BEGIN(); {\
exec; \
@ -107,10 +110,9 @@ static void reset_unblock_function(rb_thread_t *th, const struct rb_unblock_call
thread_debug("leave blocking region (%p)\n", __th); \
remove_signal_thread_list(__th); \
reset_unblock_function(__th, &__oldubf); \
if (__th->status == THREAD_STOPPED) { \
if (stopped && __th->status == THREAD_STOPPED) { \
__th->status = __prev_status; \
} \
RUBY_VM_CHECK_INTS(); \
} while(0)
#if THREAD_DEBUG
@ -263,6 +265,11 @@ rb_thread_terminate_all(void)
rb_bug("rb_thread_terminate_all: called by child thread (%p, %p)", vm->main_thread, th);
}
/* unlock all locking mutexes */
if (th->keeping_mutexes) {
rb_mutex_unlock_all(th->keeping_mutexes);
}
thread_debug("rb_thread_terminate_all (main thread: %p)\n", th);
st_foreach(vm->living_threads, terminate_i, (st_data_t)th);
@ -361,6 +368,18 @@ thread_start_func_2(rb_thread_t *th, VALUE *stack_start, VALUE *register_stack_s
}
TH_POP_TAG();
/* locking_mutex must be Qfalse */
if (th->locking_mutex != Qfalse) {
rb_bug("thread_start_func_2: locking_mutex must be NULL (%p:%ld)", th, th->locking_mutex);
}
/* unlock all locking mutexes */
if (th->keeping_mutexes) {
rb_mutex_unlock_all(th->keeping_mutexes);
th->keeping_mutexes = Qfalse;
}
/* delete self from living_threads */
st_delete_wrap(th->vm->living_threads, th->self);
/* wake up joinning threads */
@ -371,6 +390,7 @@ thread_start_func_2(rb_thread_t *th, VALUE *stack_start, VALUE *register_stack_s
join_th = join_th->join_list_next;
}
st_delete_wrap(th->vm->living_threads, th->self);
if (th != main_th) rb_check_deadlock(th->vm);
if (!th->root_fiber) {
rb_thread_recycle_stack_release(th->stack);
@ -511,7 +531,7 @@ thread_join_sleep(VALUE arg)
while (target_th->status != THREAD_KILLED) {
if (p->forever) {
sleep_forever(th);
sleep_forever(th, 1);
}
else {
now = timeofday();
@ -667,22 +687,29 @@ double2timeval(double d)
}
static void
sleep_forever(rb_thread_t *th)
sleep_forever(rb_thread_t *th, int deadlockable)
{
native_sleep(th, 0);
native_sleep(th, 0, deadlockable);
}
static void
sleep_timeval(rb_thread_t *th, struct timeval tv)
{
native_sleep(th, &tv);
native_sleep(th, &tv, 0);
}
void
rb_thread_sleep_forever()
{
thread_debug("rb_thread_sleep_forever\n");
sleep_forever(GET_THREAD());
sleep_forever(GET_THREAD(), 0);
}
static void
rb_thread_sleep_deadly()
{
thread_debug("rb_thread_sleep_deadly\n");
sleep_forever(GET_THREAD(), 1);
}
static double
@ -782,7 +809,8 @@ rb_thread_blocking_region(
BLOCKING_REGION({
val = func(data1);
}, ubf, data2);
}, ubf, data2, 1);
RUBY_VM_CHECK_INTS();
return val;
}
@ -1128,7 +1156,7 @@ rb_thread_stop(void)
rb_raise(rb_eThreadError,
"stopping only thread\n\tnote: use sleep to stop forever");
}
rb_thread_sleep_forever();
rb_thread_sleep_deadly();
return Qnil;
}
@ -1142,6 +1170,7 @@ thread_list_i(st_data_t key, st_data_t val, void *data)
switch (th->status) {
case THREAD_RUNNABLE:
case THREAD_STOPPED:
case THREAD_STOPPED_FOREVER:
case THREAD_TO_KILL:
rb_ary_push(ary, th->self);
default:
@ -1336,6 +1365,7 @@ thread_status_name(enum rb_thread_status status)
case THREAD_RUNNABLE:
return "run";
case THREAD_STOPPED:
case THREAD_STOPPED_FOREVER:
return "sleep";
case THREAD_TO_KILL:
return "aborting";
@ -1435,7 +1465,7 @@ rb_thread_stop_p(VALUE thread)
if (rb_thread_dead(th))
return Qtrue;
if (th->status == THREAD_STOPPED)
if (th->status == THREAD_STOPPED || th->status == THREAD_STOPPED_FOREVER)
return Qtrue;
return Qfalse;
}
@ -1875,14 +1905,16 @@ do_select(int n, fd_set *read, fd_set *write, fd_set *except,
if (except) *except = orig_except;
wait = &wait_100ms;
} while (__th->interrupt_flag == 0 && (timeout == 0 || subst(timeout, &wait_100ms)));
}, 0, 0);
}, 0, 0, 1);
RUBY_VM_CHECK_INTS();
} while (result == 0 && (timeout == 0 || subst(timeout, &wait_100ms)));
}
#else
BLOCKING_REGION({
result = select(n, read, write, except, timeout);
if (result < 0) lerrno = errno;
}, ubf_select, GET_THREAD());
}, ubf_select, GET_THREAD(), 1);
RUBY_VM_CHECK_INTS();
#endif
errno = lerrno;
@ -2319,12 +2351,15 @@ typedef struct mutex_struct {
rb_thread_lock_t lock;
rb_thread_cond_t cond;
rb_thread_t volatile *th;
volatile int cond_waiting;
volatile int cond_waiting, cond_notified;
VALUE next_mutex;
} mutex_t;
#define GetMutexPtr(obj, tobj) \
Data_Get_Struct(obj, mutex_t, tobj)
static const char *mutex_unlock(mutex_t *mutex);
static void
mutex_mark(void *ptr)
{
@ -2341,6 +2376,10 @@ mutex_free(void *ptr)
{
if (ptr) {
mutex_t *mutex = ptr;
if (mutex->th) {
/* rb_warn("free locked mutex"); */
mutex_unlock(mutex);
}
native_mutex_destroy(&mutex->lock);
native_cond_destroy(&mutex->cond);
}
@ -2391,6 +2430,17 @@ rb_mutex_locked_p(VALUE self)
return mutex->th ? Qtrue : Qfalse;
}
static void
mutex_locked(rb_thread_t *th, VALUE self)
{
if (th->keeping_mutexes) {
mutex_t *mutex;
GetMutexPtr(self, mutex);
mutex->next_mutex = th->keeping_mutexes;
}
th->keeping_mutexes = self;
}
/*
* call-seq:
* mutex.try_lock => true or false
@ -2413,6 +2463,8 @@ rb_mutex_trylock(VALUE self)
if (mutex->th == 0) {
mutex->th = GET_THREAD();
locked = Qtrue;
mutex_locked(GET_THREAD(), self);
}
native_mutex_unlock(&mutex->lock);
@ -2420,17 +2472,23 @@ rb_mutex_trylock(VALUE self)
}
static int
lock_func(rb_thread_t *th, mutex_t *mutex)
lock_func(rb_thread_t *th, mutex_t *mutex, int last_thread)
{
int interrupted = Qfalse;
int interrupted = 0;
native_mutex_lock(&mutex->lock);
while (mutex->th || (mutex->th = th, 0)) {
if (last_thread) {
interrupted = 2;
break;
}
mutex->cond_waiting++;
native_cond_wait(&mutex->cond, &mutex->lock);
mutex->cond_notified--;
if (th->interrupt_flag) {
interrupted = Qtrue;
if (RUBY_VM_INTERRUPTED(th)) {
interrupted = 1;
break;
}
}
@ -2445,6 +2503,7 @@ lock_interrupt(void *ptr)
native_mutex_lock(&mutex->lock);
if (mutex->cond_waiting > 0) {
native_cond_broadcast(&mutex->cond);
mutex->cond_notified = mutex->cond_waiting;
mutex->cond_waiting = 0;
}
native_mutex_unlock(&mutex->lock);
@ -2467,10 +2526,29 @@ rb_mutex_lock(VALUE self)
while (mutex->th != th) {
int interrupted;
int prev_status = th->status;
int last_thread = 0;
th->locking_mutex = self;
th->status = THREAD_STOPPED_FOREVER;
th->vm->sleeper++;
if (th->vm->living_threads->num_entries == th->vm->sleeper) {
last_thread = 1;
}
BLOCKING_REGION({
interrupted = lock_func(th, mutex);
}, lock_interrupt, mutex);
interrupted = lock_func(th, mutex, last_thread);
}, lock_interrupt, mutex, 0);
th->locking_mutex = Qfalse;
if (interrupted == 2) {
rb_check_deadlock(th->vm);
RUBY_VM_SET_TIMER_INTERRUPT(th);
}
th->status = prev_status;
th->vm->sleeper--;
if (mutex->th == th) mutex_locked(th, self);
if (interrupted) {
RUBY_VM_CHECK_INTS();
@ -2480,19 +2558,12 @@ rb_mutex_lock(VALUE self)
return self;
}
/*
* call-seq:
* mutex.unlock => self
*
* Releases the lock.
* Raises +ThreadError+ if +mutex+ wasn't locked by the current thread.
*/
VALUE
rb_mutex_unlock(VALUE self)
static const char *
mutex_unlock(mutex_t *mutex)
{
mutex_t *mutex;
const char *err = NULL;
GetMutexPtr(self, mutex);
rb_thread_t *th = GET_THREAD();
mutex_t *th_mutex;
native_mutex_lock(&mutex->lock);
@ -2513,15 +2584,68 @@ rb_mutex_unlock(VALUE self)
native_mutex_unlock(&mutex->lock);
if (!err) {
GetMutexPtr(th->keeping_mutexes, th_mutex);
if (th_mutex == mutex) {
th->keeping_mutexes = mutex->next_mutex;
}
else {
while (1) {
mutex_t *tmp_mutex;
GetMutexPtr(th_mutex->next_mutex, tmp_mutex);
if (tmp_mutex == mutex) {
th_mutex->next_mutex = tmp_mutex->next_mutex;
break;
}
th_mutex = tmp_mutex;
}
}
mutex->next_mutex = Qfalse;
}
return err;
}
/*
* call-seq:
* mutex.unlock => self
*
* Releases the lock.
* Raises +ThreadError+ if +mutex+ wasn't locked by the current thread.
*/
VALUE
rb_mutex_unlock(VALUE self)
{
const char *err;
mutex_t *mutex;
GetMutexPtr(self, mutex);
err = mutex_unlock(mutex);
if (err) rb_raise(rb_eThreadError, err);
return self;
}
static void
rb_mutex_unlock_all(VALUE mutexes)
{
const char *err;
mutex_t *mutex;
while (mutexes) {
GetMutexPtr(mutexes, mutex);
/* rb_warn("mutex #<%s:%p> remains to be locked by terminated thread",
rb_obj_classname(mutexes), (void*)mutexes); */
mutexes = mutex->next_mutex;
err = mutex_unlock(mutex);
if (err) rb_bug("invalid keeping_mutexes");
}
}
static VALUE
rb_mutex_sleep_forever(VALUE time)
{
rb_thread_sleep_forever();
rb_thread_sleep_deadly();
return Qnil;
}
@ -3275,3 +3399,44 @@ ruby_native_thread_p(void)
return th ? Qtrue : Qfalse;
}
static int
check_deadlock_i(st_data_t key, st_data_t val, int *found)
{
VALUE thval = key;
rb_thread_t *th;
GetThreadPtr(thval, th);
if (th->status != THREAD_STOPPED_FOREVER || RUBY_VM_INTERRUPTED(th)) {
*found = 1;
}
else if (th->locking_mutex) {
mutex_t *mutex;
GetMutexPtr(th->locking_mutex, mutex);
native_mutex_lock(&mutex->lock);
if (mutex->th == th || (!mutex->th && mutex->cond_notified)) {
*found = 1;
}
native_mutex_unlock(&mutex->lock);
}
return (*found) ? ST_STOP : ST_CONTINUE;
}
static void
rb_check_deadlock(rb_vm_t *vm)
{
int found = 0;
if (vm->living_threads->num_entries != vm->sleeper) return;
st_foreach(vm->living_threads, check_deadlock_i, (st_data_t)&found);
if (!found) {
VALUE argv[2];
argv[0] = rb_eFatal;
argv[1] = rb_str_new2("deadlock detected");
rb_thread_raise(2, argv, vm->main_thread);
}
}

View file

@ -402,7 +402,7 @@ ubf_select(void *ptr)
#endif
static void
native_sleep(rb_thread_t *th, struct timeval *tv)
native_sleep(rb_thread_t *th, struct timeval *tv, int deadlockable)
{
int prev_status = th->status;
struct timespec ts;
@ -418,7 +418,14 @@ native_sleep(rb_thread_t *th, struct timeval *tv)
}
}
if (!tv && deadlockable) {
th->status = THREAD_STOPPED_FOREVER;
th->vm->sleeper++;
rb_check_deadlock(th->vm);
}
else {
th->status = THREAD_STOPPED;
}
thread_debug("native_sleep %ld\n", tv ? tv->tv_sec : -1);
GVL_UNLOCK_BEGIN();
@ -455,9 +462,10 @@ native_sleep(rb_thread_t *th, struct timeval *tv)
th->unblock.arg = 0;
pthread_mutex_unlock(&th->interrupt_lock);
th->status = prev_status;
}
GVL_UNLOCK_END();
th->status = prev_status;
if (!tv && deadlockable) th->vm->sleeper--;
RUBY_VM_CHECK_INTS();
thread_debug("native_sleep done\n");

View file

@ -204,7 +204,7 @@ rb_w32_Sleep(unsigned long msec)
}
static void
native_sleep(rb_thread_t *th, struct timeval *tv)
native_sleep(rb_thread_t *th, struct timeval *tv, int deadlockable)
{
DWORD msec;
if (tv) {
@ -214,12 +214,19 @@ native_sleep(rb_thread_t *th, struct timeval *tv)
msec = INFINITE;
}
if (!tv && deadlockable) {
th->status = THREAD_STOPPED_FOREVER;
th->vm->sleeper++;
rb_check_deadlock(th->vm);
}
else {
th->status = THREAD_STOPPED;
}
GVL_UNLOCK_BEGIN();
{
DWORD ret;
int status = th->status;
th->status = THREAD_STOPPED;
th->unblock.func = ubf_handle;
th->unblock.arg = th;
@ -234,9 +241,10 @@ native_sleep(rb_thread_t *th, struct timeval *tv)
th->unblock.func = 0;
th->unblock.arg = 0;
th->status = status;
}
GVL_UNLOCK_END();
th->status = status;
if (!tv && deadlockable) th->vm->sleeper++;
RUBY_VM_CHECK_INTS();
}

9
vm.c
View file

@ -1470,6 +1470,13 @@ thread_free(void *ptr)
RUBY_FREE_UNLESS_NULL(th->stack);
}
if (th->locking_mutex != Qfalse) {
rb_bug("thread_free: locking_mutex must be NULL (%p:%ld)", th, th->locking_mutex);
}
if (th->keeping_mutexes != Qfalse) {
rb_bug("thread_free: keeping_mutexes must be NULL (%p:%ld)", th, th->locking_mutex);
}
if (th->local_storage) {
st_free_table(th->local_storage);
}
@ -1537,6 +1544,8 @@ rb_thread_mark(void *ptr)
RUBY_MARK_UNLESS_NULL(th->root_fiber);
RUBY_MARK_UNLESS_NULL(th->stat_insn_usage);
RUBY_MARK_UNLESS_NULL(th->locking_mutex);
rb_mark_tbl(th->local_storage);
if (GET_THREAD() != th && th->machine_stack_start && th->machine_stack_end) {

View file

@ -302,6 +302,7 @@ struct rb_vm_struct
int running;
int thread_abort_on_exception;
unsigned long trace_flag;
volatile int sleeper;
/* object management */
VALUE mark_object_ary;
@ -360,6 +361,7 @@ enum rb_thread_status {
THREAD_TO_KILL,
THREAD_RUNNABLE,
THREAD_STOPPED,
THREAD_STOPPED_FOREVER,
THREAD_KILLED,
};
@ -429,6 +431,8 @@ struct rb_thread_struct
int interrupt_flag;
rb_thread_lock_t interrupt_lock;
struct rb_unblock_callback unblock;
VALUE locking_mutex;
VALUE keeping_mutexes;
struct rb_vm_tag *tag;
struct rb_vm_trap_tag *trap_tag;