/* -*-c-*- */ /********************************************************************** thread_win32.c - $Author: yugui $ Copyright (C) 2004-2007 Koichi Sasada **********************************************************************/ #ifdef THREAD_SYSTEM_DEPENDENT_IMPLEMENTATION #include //#define THREAD_DEBUG 1 #define WIN32_WAIT_TIMEOUT 10 /* 10 ms */ #undef Sleep #define native_thread_yield() Sleep(0) #define remove_signal_thread_list(th) static volatile DWORD ruby_native_thread_key = TLS_OUT_OF_INDEXES; static int native_mutex_lock(rb_thread_lock_t *); static int native_mutex_unlock(rb_thread_lock_t *); static int native_mutex_trylock(rb_thread_lock_t *); void native_mutex_initialize(rb_thread_lock_t *); static void native_cond_signal(rb_thread_cond_t *cond); static void native_cond_broadcast(rb_thread_cond_t *cond); static void native_cond_wait(rb_thread_cond_t *cond, rb_thread_lock_t *mutex); static void native_cond_initialize(rb_thread_cond_t *cond); void native_cond_destroy(rb_thread_cond_t *cond); static rb_thread_t * ruby_thread_from_native(void) { return TlsGetValue(ruby_native_thread_key); } static int ruby_thread_set_native(rb_thread_t *th) { return TlsSetValue(ruby_native_thread_key, th); } static void Init_native_thread(void) { rb_thread_t *th = GET_THREAD(); ruby_native_thread_key = TlsAlloc(); ruby_thread_set_native(th); DuplicateHandle(GetCurrentProcess(), GetCurrentThread(), GetCurrentProcess(), &th->thread_id, 0, FALSE, DUPLICATE_SAME_ACCESS); th->native_thread_data.interrupt_event = CreateEvent(0, TRUE, FALSE, 0); thread_debug("initial thread (th: %p, thid: %p, event: %p)\n", th, GET_THREAD()->thread_id, th->native_thread_data.interrupt_event); } static void w32_error(void) { LPVOID lpMsgBuf = 0; int nError = GetLastError(); FormatMessage(FORMAT_MESSAGE_ALLOCATE_BUFFER | FORMAT_MESSAGE_FROM_SYSTEM | FORMAT_MESSAGE_IGNORE_INSERTS, NULL, nError, MAKELANGID(LANG_NEUTRAL, SUBLANG_DEFAULT), (LPTSTR) & lpMsgBuf, 0, NULL); if ( lpMsgBuf ) rb_bug("%s;Error code: %d", (char*)lpMsgBuf, nError); else rb_bug("Error code: %d", nError); } static void w32_set_event(HANDLE handle) { if (SetEvent(handle) == 0) { w32_error(); } } static void w32_reset_event(HANDLE handle) { if (ResetEvent(handle) == 0) { w32_error(); } } static int w32_wait_events(HANDLE *events, int count, DWORD timeout, rb_thread_t *th) { HANDLE *targets = events; HANDLE intr; DWORD ret; thread_debug(" w32_wait_events events:%p, count:%d, timeout:%ld, th:%p\n", events, count, timeout, th); if (th && (intr = th->native_thread_data.interrupt_event)) { w32_reset_event(intr); if (RUBY_VM_INTERRUPTED(th)) { w32_set_event(intr); } targets = ALLOCA_N(HANDLE, count + 1); memcpy(targets, events, sizeof(HANDLE) * count); targets[count++] = intr; thread_debug(" * handle: %p (count: %d, intr)\n", intr, count); } thread_debug(" WaitForMultipleObjects start (count: %d)\n", count); ret = WaitForMultipleObjects(count, targets, FALSE, timeout); thread_debug(" WaitForMultipleObjects end (ret: %lu)\n", ret); if (ret == WAIT_OBJECT_0 + count - 1 && th) { errno = EINTR; } if (ret == -1 && THREAD_DEBUG) { int i; DWORD dmy; for (i = 0; i < count; i++) { thread_debug(" * error handle %d - %s\n", i, GetHandleInformation(targets[i], &dmy) ? "OK" : "NG"); } } return ret; } static void ubf_handle(void *ptr); #define ubf_select ubf_handle int rb_w32_wait_events_blocking(HANDLE *events, int num, DWORD timeout) { return w32_wait_events(events, num, timeout, GET_THREAD()); } int rb_w32_wait_events(HANDLE *events, int num, DWORD timeout) { int ret; BLOCKING_REGION(ret = rb_w32_wait_events_blocking(events, num, timeout), ubf_handle, GET_THREAD()); return ret; } static void w32_close_handle(HANDLE handle) { if (CloseHandle(handle) == 0) { w32_error(); } } static void w32_resume_thread(HANDLE handle) { if (ResumeThread(handle) == -1) { w32_error(); } } #if defined _MSC_VER && !defined(_WIN32_WCE) #define HAVE__BEGINTHREADEX 1 #else #undef HAVE__BEGINTHREADEX #endif #ifdef HAVE__BEGINTHREADEX #define start_thread (HANDLE)_beginthreadex typedef unsigned long (_stdcall *w32_thread_start_func)(void*); #else #define start_thread CreateThread typedef LPTHREAD_START_ROUTINE w32_thread_start_func; #endif static HANDLE w32_create_thread(DWORD stack_size, w32_thread_start_func func, void *val) { return start_thread(0, stack_size, func, val, CREATE_SUSPENDED #ifdef _WIN32_WCE |STACK_SIZE_PARAM_IS_A_RESERVATION #endif ,0); } int rb_w32_sleep(unsigned long msec) { return w32_wait_events(0, 0, msec, GET_THREAD()); } int WINAPI rb_w32_Sleep(unsigned long msec) { int ret; BLOCKING_REGION(ret = rb_w32_sleep(msec), ubf_handle, GET_THREAD()); return ret; } static void native_sleep(rb_thread_t *th, struct timeval *tv) { DWORD msec; if (tv) { msec = tv->tv_sec * 1000 + tv->tv_usec / 1000; } else { msec = INFINITE; } GVL_UNLOCK_BEGIN(); { DWORD ret; native_mutex_lock(&th->interrupt_lock); th->unblock.func = ubf_handle; th->unblock.arg = th; native_mutex_unlock(&th->interrupt_lock); if (RUBY_VM_INTERRUPTED(th)) { /* interrupted. return immediate */ } else { thread_debug("native_sleep start (%lu)\n", msec); ret = w32_wait_events(0, 0, msec, th); thread_debug("native_sleep done (%lu)\n", ret); } native_mutex_lock(&th->interrupt_lock); th->unblock.func = 0; th->unblock.arg = 0; native_mutex_unlock(&th->interrupt_lock); } GVL_UNLOCK_END(); } static int native_mutex_lock(rb_thread_lock_t *lock) { #if USE_WIN32_MUTEX DWORD result; while (1) { thread_debug("native_mutex_lock: %p\n", *lock); result = w32_wait_events(&*lock, 1, INFINITE, 0); switch (result) { case WAIT_OBJECT_0: /* get mutex object */ thread_debug("acquire mutex: %p\n", *lock); return 0; case WAIT_OBJECT_0 + 1: /* interrupt */ errno = EINTR; thread_debug("acquire mutex interrupted: %p\n", *lock); return 0; case WAIT_TIMEOUT: thread_debug("timeout mutex: %p\n", *lock); break; case WAIT_ABANDONED: rb_bug("win32_mutex_lock: WAIT_ABANDONED"); break; default: rb_bug("win32_mutex_lock: unknown result (%d)", result); break; } } return 0; #else EnterCriticalSection(lock); return 0; #endif } static int native_mutex_unlock(rb_thread_lock_t *lock) { #if USE_WIN32_MUTEX thread_debug("release mutex: %p\n", *lock); return ReleaseMutex(*lock); #else LeaveCriticalSection(lock); return 0; #endif } static int native_mutex_trylock(rb_thread_lock_t *lock) { #if USE_WIN32_MUTEX int result; thread_debug("native_mutex_trylock: %p\n", *lock); result = w32_wait_events(&*lock, 1, 1, 0); thread_debug("native_mutex_trylock result: %d\n", result); switch (result) { case WAIT_OBJECT_0: return 0; case WAIT_TIMEOUT: return EBUSY; } return EINVAL; #else return TryEnterCriticalSection(lock) == 0; #endif } void native_mutex_initialize(rb_thread_lock_t *lock) { #if USE_WIN32_MUTEX *lock = CreateMutex(NULL, FALSE, NULL); if (*lock == NULL) { w32_error(); } /* thread_debug("initialize mutex: %p\n", *lock); */ #else InitializeCriticalSection(lock); #endif } void native_mutex_destroy(rb_thread_lock_t *lock) { #if USE_WIN32_MUTEX w32_close_handle(lock); #else DeleteCriticalSection(lock); #endif } struct cond_event_entry { struct cond_event_entry* next; HANDLE event; }; struct rb_thread_cond_struct { struct cond_event_entry *next; struct cond_event_entry *last; }; static void native_cond_signal(rb_thread_cond_t *cond) { /* cond is guarded by mutex */ struct cond_event_entry *e = cond->next; if (e) { cond->next = e->next; SetEvent(e->event); } else { rb_bug("native_cond_signal: no pending threads"); } } static void native_cond_broadcast(rb_thread_cond_t *cond) { /* cond is guarded by mutex */ struct cond_event_entry *e = cond->next; cond->next = 0; while (e) { SetEvent(e->event); e = e->next; } } static void native_cond_wait(rb_thread_cond_t *cond, rb_thread_lock_t *mutex) { DWORD r; struct cond_event_entry entry; entry.next = 0; entry.event = CreateEvent(0, FALSE, FALSE, 0); /* cond is guarded by mutex */ if (cond->next) { cond->last->next = &entry; cond->last = &entry; } else { cond->next = &entry; cond->last = &entry; } native_mutex_unlock(mutex); { r = WaitForSingleObject(entry.event, INFINITE); if (r != WAIT_OBJECT_0) { rb_bug("native_cond_wait: WaitForSingleObject returns %lu", r); } } native_mutex_lock(mutex); w32_close_handle(entry.event); } static void native_cond_initialize(rb_thread_cond_t *cond) { cond->next = 0; cond->last = 0; } static void native_cond_destroy(rb_thread_cond_t *cond) { /* */ } void ruby_init_stack(VALUE *addr) { } #define CHECK_ERR(expr) \ {if (!(expr)) {rb_bug("err: %lu - %s", GetLastError(), #expr);}} static void native_thread_init_stack(rb_thread_t *th) { MEMORY_BASIC_INFORMATION mi; char *base, *end; DWORD size, space; CHECK_ERR(VirtualQuery(&mi, &mi, sizeof(mi))); base = mi.AllocationBase; end = mi.BaseAddress; end += mi.RegionSize; size = end - base; space = size / 5; if (space > 1024*1024) space = 1024*1024; th->machine_stack_start = (VALUE *)end - 1; th->machine_stack_maxsize = size - space; } static void native_thread_destroy(rb_thread_t *th) { HANDLE intr = th->native_thread_data.interrupt_event; native_mutex_destroy(&th->interrupt_lock); thread_debug("close handle - intr: %p, thid: %p\n", intr, th->thread_id); th->native_thread_data.interrupt_event = 0; if ( intr ) w32_close_handle(intr); } static unsigned long _stdcall thread_start_func_1(void *th_ptr) { rb_thread_t *th = th_ptr; volatile HANDLE thread_id = th->thread_id; native_thread_init_stack(th); /* th->native_thread_data.interrupt_event = CreateEvent(0, TRUE, FALSE, 0); if ( !th->native_thread_data.interrupt_event) { DWORD dwErr = GetLastError(); thread_debug("thread interrupt_event error: %d\n", dwErr ); } */ /* run */ thread_debug("thread created (th: %p, thid: %p, event: %p)\n", th, th->thread_id, th->native_thread_data.interrupt_event); thread_start_func_2(th, th->machine_stack_start, rb_ia64_bsp()); w32_close_handle(thread_id); thread_debug("thread deleted (th: %p)\n", th); return 0; } static int native_thread_create(rb_thread_t *th) { size_t stack_size = 4 * 1024; /* 4KB */ th->native_thread_data.interrupt_event = CreateEvent(0, TRUE, FALSE, 0); if ( !th->native_thread_data.interrupt_event) { DWORD dwErr = GetLastError(); thread_debug("thread interrupt_event error: %d\n", dwErr ); } th->thread_id = w32_create_thread(stack_size, thread_start_func_1, th); if ((th->thread_id) == 0) { st_delete_wrap(th->vm->living_threads, th->self); if ( th->native_thread_data.interrupt_event ) w32_close_handle(th->native_thread_data.interrupt_event); rb_raise(rb_eThreadError, "can't create Thread (%d)", errno); } w32_resume_thread(th->thread_id); if (THREAD_DEBUG) { Sleep(0); thread_debug("create: (th: %p, thid: %p, intr: %p), stack size: %d\n", th, th->thread_id, th->native_thread_data.interrupt_event, stack_size); } return 0; } static void native_thread_join(HANDLE th) { w32_wait_events(&th, 1, 0, 0); } #if USE_NATIVE_THREAD_PRIORITY static void native_thread_apply_priority(rb_thread_t *th) { int priority = th->priority; if (th->priority > 0) { priority = THREAD_PRIORITY_ABOVE_NORMAL; } else if (th->priority < 0) { priority = THREAD_PRIORITY_BELOW_NORMAL; } else { priority = THREAD_PRIORITY_NORMAL; } SetThreadPriority(th->thread_id, priority); } #endif /* USE_NATIVE_THREAD_PRIORITY */ static void ubf_handle(void *ptr) { typedef BOOL (WINAPI *cancel_io_func_t)(HANDLE); rb_thread_t *th = (rb_thread_t *)ptr; thread_debug("ubf_handle: %p\n", th); w32_set_event(th->native_thread_data.interrupt_event); } static HANDLE timer_thread_id = 0; static HANDLE timer_thread_lock; static unsigned long _stdcall timer_thread_func(void *dummy) { thread_debug("timer_thread\n"); while (WaitForSingleObject(timer_thread_lock, WIN32_WAIT_TIMEOUT) == WAIT_TIMEOUT) { timer_thread_function(dummy); } thread_debug("timer killed\n"); return 0; } static void rb_thread_create_timer_thread(void) { if (timer_thread_id == 0) { if (!timer_thread_lock) { timer_thread_lock = CreateEvent(0, TRUE, FALSE, 0); } timer_thread_id = w32_create_thread(1024 + (THREAD_DEBUG ? BUFSIZ : 0), timer_thread_func, 0); w32_resume_thread(timer_thread_id); } } static int native_stop_timer_thread(void) { int stopped = --system_working <= 0; if (stopped) { CloseHandle(timer_thread_lock); timer_thread_lock = 0; } return stopped; } #endif /* THREAD_SYSTEM_DEPENDENT_IMPLEMENTATION */