12#ifdef THREAD_SYSTEM_DEPENDENT_IMPLEMENTATION
16#define TIME_QUANTUM_USEC (10 * 1000)
17#define RB_CONDATTR_CLOCK_MONOTONIC 1
21#define native_thread_yield() Sleep(0)
22#define unregister_ubf_list(th)
23#define ubf_wakeup_all_threads() do {} while (0)
24#define ubf_threads_empty() (1)
25#define ubf_timer_disarm() do {} while (0)
26#define ubf_list_atfork() do {} while (0)
28static volatile DWORD ruby_native_thread_key = TLS_OUT_OF_INDEXES;
30static int w32_wait_events(HANDLE *events,
int count, DWORD timeout,
rb_thread_t *th);
32rb_internal_thread_event_hook_t *
48w32_error(const
char *func)
51 DWORD err = GetLastError();
52 if (FormatMessage(FORMAT_MESSAGE_ALLOCATE_BUFFER |
53 FORMAT_MESSAGE_FROM_SYSTEM |
54 FORMAT_MESSAGE_IGNORE_INSERTS,
57 MAKELANGID(LANG_ENGLISH, SUBLANG_ENGLISH_US),
58 (LPTSTR) & lpMsgBuf, 0, NULL) == 0)
59 FormatMessage(FORMAT_MESSAGE_ALLOCATE_BUFFER |
60 FORMAT_MESSAGE_FROM_SYSTEM |
61 FORMAT_MESSAGE_IGNORE_INSERTS,
64 MAKELANGID(LANG_NEUTRAL, SUBLANG_DEFAULT),
65 (LPTSTR) & lpMsgBuf, 0, NULL);
66 rb_bug(
"%s: %s", func, (
char*)lpMsgBuf);
70#define W32_EVENT_DEBUG 0
73#define w32_event_debug printf
75#define w32_event_debug if (0) printf
79w32_mutex_lock(HANDLE lock,
bool try)
84 w32_event_debug(
"lock:%p\n", lock);
86 result = w32_wait_events(&lock, 1,
try ? 0 : INFINITE, 0);
90 w32_event_debug(
"locked lock:%p\n", lock);
93 case WAIT_OBJECT_0 + 1:
96 w32_event_debug(
"interrupted lock:%p\n", lock);
100 w32_event_debug(
"timeout locK:%p\n", lock);
104 rb_bug(
"win32_mutex_lock: WAIT_ABANDONED");
108 rb_bug(
"win32_mutex_lock: unknown result (%ld)", result);
116w32_mutex_create(
void)
118 HANDLE lock = CreateMutex(NULL, FALSE, NULL);
120 w32_error(
"rb_native_mutex_initialize");
130 w32_mutex_lock(sched->lock,
false);
131 if (GVL_DEBUG) fprintf(stderr,
"gvl acquire (%p): acquire\n", th);
134#define thread_sched_to_dead thread_sched_to_waiting
139 ReleaseMutex(sched->lock);
145 thread_sched_to_waiting(sched, th);
146 native_thread_yield();
147 thread_sched_to_running(sched, th);
153 if (GVL_DEBUG) fprintf(stderr,
"sched init\n");
154 sched->lock = w32_mutex_create();
161 if (GVL_DEBUG) fprintf(stderr,
"sched destroy\n");
162 CloseHandle(sched->lock);
166ruby_thread_from_native(
void)
168 return TlsGetValue(ruby_native_thread_key);
175 rb_ractor_set_current_ec(th->ractor, th->ec);
177 return TlsSetValue(ruby_native_thread_key, th);
183 if ((ruby_current_ec_key = TlsAlloc()) == TLS_OUT_OF_INDEXES) {
184 rb_bug(
"TlsAlloc() for ruby_current_ec_key fails");
186 if ((ruby_native_thread_key = TlsAlloc()) == TLS_OUT_OF_INDEXES) {
187 rb_bug(
"TlsAlloc() for ruby_native_thread_key fails");
192 ruby_thread_set_native(main_th);
193 main_th->nt->interrupt_event = CreateEvent(0, TRUE, FALSE, 0);
195 DuplicateHandle(GetCurrentProcess(),
198 &main_th->nt->thread_id, 0, FALSE, DUPLICATE_SAME_ACCESS);
200 RUBY_DEBUG_LOG(
"initial thread th:%u thid:%p, event: %p",
201 rb_th_serial(main_th),
202 main_th->nt->thread_id,
203 main_th->nt->interrupt_event);
207ruby_mn_threads_params(
void)
212w32_wait_events(HANDLE *events,
int count, DWORD timeout,
rb_thread_t *th)
214 HANDLE *targets = events;
216 const int initcount = count;
219 w32_event_debug(
"events:%p, count:%d, timeout:%ld, th:%u\n",
220 events, count, timeout, th ? rb_th_serial(th) : UINT_MAX);
222 if (th && (intr = th->nt->interrupt_event)) {
223 if (ResetEvent(intr) && (!RUBY_VM_INTERRUPTED(th->ec) || SetEvent(intr))) {
224 targets =
ALLOCA_N(HANDLE, count + 1);
225 memcpy(targets, events,
sizeof(HANDLE) * count);
227 targets[count++] = intr;
228 w32_event_debug(
"handle:%p (count:%d, intr)\n", intr, count);
230 else if (intr == th->nt->interrupt_event) {
231 w32_error(
"w32_wait_events");
235 w32_event_debug(
"WaitForMultipleObjects start count:%d\n", count);
236 ret = WaitForMultipleObjects(count, targets, FALSE, timeout);
237 w32_event_debug(
"WaitForMultipleObjects end ret:%lu\n", ret);
239 if (ret == (DWORD)(WAIT_OBJECT_0 + initcount) && th) {
242 if (ret == WAIT_FAILED && W32_EVENT_DEBUG) {
245 for (i = 0; i < count; i++) {
246 w32_event_debug(
"i:%d %s\n", i, GetHandleInformation(targets[i], &dmy) ?
"OK" :
"NG");
252static void ubf_handle(
void *ptr);
253#define ubf_select ubf_handle
256rb_w32_wait_events_blocking(HANDLE *events,
int num, DWORD timeout)
258 return w32_wait_events(events, num, timeout, ruby_thread_from_native());
262rb_w32_wait_events(HANDLE *events,
int num, DWORD timeout)
267 BLOCKING_REGION(th, ret = rb_w32_wait_events_blocking(events, num, timeout),
268 ubf_handle, ruby_thread_from_native(), FALSE);
273w32_close_handle(HANDLE handle)
275 if (CloseHandle(handle) == 0) {
276 w32_error(
"w32_close_handle");
281w32_resume_thread(HANDLE handle)
283 if (ResumeThread(handle) == (DWORD)-1) {
284 w32_error(
"w32_resume_thread");
289#define HAVE__BEGINTHREADEX 1
291#undef HAVE__BEGINTHREADEX
294#ifdef HAVE__BEGINTHREADEX
295#define start_thread (HANDLE)_beginthreadex
296#define thread_errno errno
297typedef unsigned long (__stdcall *w32_thread_start_func)(
void*);
299#define start_thread CreateThread
300#define thread_errno rb_w32_map_errno(GetLastError())
301typedef LPTHREAD_START_ROUTINE w32_thread_start_func;
305w32_create_thread(DWORD stack_size, w32_thread_start_func func,
void *val)
307 return start_thread(0, stack_size, func, val, CREATE_SUSPENDED | STACK_SIZE_PARAM_IS_A_RESERVATION, 0);
311rb_w32_sleep(
unsigned long msec)
313 return w32_wait_events(0, 0, msec, ruby_thread_from_native());
317rb_w32_Sleep(
unsigned long msec)
322 BLOCKING_REGION(th, ret = rb_w32_sleep(msec),
323 ubf_handle, ruby_thread_from_native(), FALSE);
328hrtime2msec(rb_hrtime_t hrt)
330 return (DWORD)hrt / (DWORD)RB_HRTIME_PER_MSEC;
336 const volatile DWORD msec = rel ? hrtime2msec(*rel) : INFINITE;
338 THREAD_BLOCKING_BEGIN(th);
343 th->unblock.func = ubf_handle;
344 th->unblock.arg = th;
347 if (RUBY_VM_INTERRUPTED(th->ec)) {
351 RUBY_DEBUG_LOG(
"start msec:%lu", msec);
352 ret = w32_wait_events(0, 0, msec, th);
353 RUBY_DEBUG_LOG(
"done ret:%lu", ret);
358 th->unblock.func = 0;
362 THREAD_BLOCKING_END(th);
368#ifdef USE_WIN32_MUTEX
369 w32_mutex_lock(lock->mutex,
false);
371 EnterCriticalSection(&lock->crit);
378#ifdef USE_WIN32_MUTEX
379 return w32_mutex_lock(lock->mutex,
true);
381 return TryEnterCriticalSection(&lock->crit) == 0 ? EBUSY : 0;
388#ifdef USE_WIN32_MUTEX
389 RUBY_DEBUG_LOG(
"lock:%p", lock->mutex);
390 ReleaseMutex(lock->mutex);
392 LeaveCriticalSection(&lock->crit);
399#ifdef USE_WIN32_MUTEX
400 lock->mutex = w32_mutex_create();
403 InitializeCriticalSection(&lock->crit);
410#ifdef USE_WIN32_MUTEX
411 w32_close_handle(lock->mutex);
413 DeleteCriticalSection(&lock->crit);
417struct cond_event_entry {
418 struct cond_event_entry* next;
419 struct cond_event_entry* prev;
427 struct cond_event_entry *e = cond->next;
428 struct cond_event_entry *head = (
struct cond_event_entry*)cond;
431 struct cond_event_entry *next = e->next;
432 struct cond_event_entry *prev = e->prev;
436 e->next = e->prev = e;
446 struct cond_event_entry *e = cond->next;
447 struct cond_event_entry *head = (
struct cond_event_entry*)cond;
450 struct cond_event_entry *next = e->next;
451 struct cond_event_entry *prev = e->prev;
457 e->next = e->prev = e;
464native_cond_timedwait_ms(rb_nativethread_cond_t *cond, rb_nativethread_lock_t *mutex,
unsigned long msec)
467 struct cond_event_entry entry;
468 struct cond_event_entry *head = (
struct cond_event_entry*)cond;
470 entry.event = CreateEvent(0, FALSE, FALSE, 0);
474 entry.prev = head->prev;
475 head->prev->next = &entry;
480 r = WaitForSingleObject(entry.event, msec);
481 if ((r != WAIT_OBJECT_0) && (r != WAIT_TIMEOUT)) {
482 rb_bug(
"rb_native_cond_wait: WaitForSingleObject returns %lu", r);
487 entry.prev->next = entry.next;
488 entry.next->prev = entry.prev;
490 w32_close_handle(entry.event);
491 return (r == WAIT_OBJECT_0) ? 0 : ETIMEDOUT;
497 native_cond_timedwait_ms(cond, mutex, INFINITE);
501abs_timespec_to_timeout_ms(
const struct timespec *ts)
506 gettimeofday(&now, NULL);
507 tv.tv_sec = ts->tv_sec;
508 tv.tv_usec = ts->tv_nsec / 1000;
510 if (!rb_w32_time_subtract(&tv, &now))
513 return (tv.tv_sec * 1000) + (tv.tv_usec / 1000);
517native_cond_timedwait(rb_nativethread_cond_t *cond, rb_nativethread_lock_t *mutex,
const struct timespec *ts)
519 unsigned long timeout_ms;
521 timeout_ms = abs_timespec_to_timeout_ms(ts);
525 return native_cond_timedwait_ms(cond, mutex, timeout_ms);
528static struct timespec native_cond_timeout(rb_nativethread_cond_t *cond,
struct timespec timeout_rel);
534 .tv_sec = msec / 1000,
535 .tv_nsec = (msec % 1000) * 1000 * 1000,
537 struct timespec ts = native_cond_timeout(cond, rel);
538 native_cond_timedwait(cond, mutex, &ts);
542native_cond_timeout(rb_nativethread_cond_t *cond,
struct timespec timeout_rel)
549 ret = gettimeofday(&tv, 0);
552 now.tv_sec = tv.tv_sec;
553 now.tv_nsec = tv.tv_usec * 1000;
555 timeout.tv_sec = now.tv_sec;
556 timeout.tv_nsec = now.tv_nsec;
557 timeout.tv_sec += timeout_rel.tv_sec;
558 timeout.tv_nsec += timeout_rel.tv_nsec;
560 if (timeout.tv_nsec >= 1000*1000*1000) {
562 timeout.tv_nsec -= 1000*1000*1000;
565 if (timeout.tv_sec < now.tv_sec)
566 timeout.tv_sec = TIMET_MAX;
574 cond->next = (
struct cond_event_entry *)cond;
575 cond->prev = (
struct cond_event_entry *)cond;
589#define CHECK_ERR(expr) \
590 {if (!(expr)) {rb_bug("err: %lu - %s", GetLastError(), #expr);}}
594COMPILER_WARNING_IGNORED(-Wmaybe-uninitialized)
597query_memory_basic_info(PMEMORY_BASIC_INFORMATION mi)
599 return VirtualQuery(mi, mi,
sizeof(*mi));
606 MEMORY_BASIC_INFORMATION mi;
610 CHECK_ERR(query_memory_basic_info(&mi));
611 base = mi.AllocationBase;
612 end = mi.BaseAddress;
613 end += mi.RegionSize;
616 if (space > 1024*1024) space = 1024*1024;
617 th->ec->machine.stack_start = (
VALUE *)end - 1;
618 th->ec->machine.stack_maxsize = size - space;
621#ifndef InterlockedExchangePointer
622#define InterlockedExchangePointer(t, v) \
623 (void *)InterlockedExchange((long *)(t), (long)(v))
629 HANDLE intr = InterlockedExchangePointer(&nt->interrupt_event, 0);
630 RUBY_DEBUG_LOG(
"close handle intr:%p, thid:%p\n", intr, nt->thread_id);
631 w32_close_handle(intr);
635static unsigned long __stdcall
636thread_start_func_1(
void *th_ptr)
639 volatile HANDLE thread_id = th->nt->thread_id;
641 native_thread_init_stack(th);
642 th->nt->interrupt_event = CreateEvent(0, TRUE, FALSE, 0);
645 RUBY_DEBUG_LOG(
"thread created th:%u, thid: %p, event: %p",
646 rb_th_serial(th), th->nt->thread_id, th->nt->interrupt_event);
648 thread_sched_to_running(TH_SCHED(th), th);
649 ruby_thread_set_native(th);
652 thread_start_func_2(th, th->ec->machine.stack_start);
654 w32_close_handle(thread_id);
655 RUBY_DEBUG_LOG(
"thread deleted th:%u", rb_th_serial(th));
664 const size_t stack_size = th->vm->default_params.thread_machine_stack_size;
666 th->nt->thread_id = w32_create_thread(stack_size, thread_start_func_1, th);
669 size_t vm_stack_word_size = th->vm->default_params.thread_vm_stack_size /
sizeof(
VALUE);
670 void *vm_stack = ruby_xmalloc(vm_stack_word_size *
sizeof(
VALUE));
671 th->sched.vm_stack = vm_stack;
672 rb_ec_initialize_vm_stack(th->ec, vm_stack, vm_stack_word_size);
674 if ((th->nt->thread_id) == 0) {
678 w32_resume_thread(th->nt->thread_id);
680 if (USE_RUBY_DEBUG_LOG) {
682 RUBY_DEBUG_LOG(
"th:%u thid:%p intr:%p), stack size: %"PRIuSIZE
"",
683 rb_th_serial(th), th->nt->thread_id,
684 th->nt->interrupt_event, stack_size);
690native_thread_join(HANDLE th)
692 w32_wait_events(&th, 1, INFINITE, 0);
695#if USE_NATIVE_THREAD_PRIORITY
700 int priority = th->priority;
701 if (th->priority > 0) {
702 priority = THREAD_PRIORITY_ABOVE_NORMAL;
704 else if (th->priority < 0) {
705 priority = THREAD_PRIORITY_BELOW_NORMAL;
708 priority = THREAD_PRIORITY_NORMAL;
711 SetThreadPriority(th->nt->thread_id, priority);
716int rb_w32_select_with_thread(
int, fd_set *, fd_set *, fd_set *,
struct timeval *,
void *);
721 fd_set *r = NULL, *w = NULL, *e = NULL;
734 return rb_w32_select_with_thread(n, r, w, e, timeout, th);
741 return w32_wait_events(0, 0, 0, th);
748 RUBY_DEBUG_LOG(
"th:%u\n", rb_th_serial(th));
750 if (!SetEvent(th->nt->interrupt_event)) {
751 w32_error(
"ubf_handle");
755int rb_w32_set_thread_description(HANDLE th,
const WCHAR *name);
756int rb_w32_set_thread_description_str(HANDLE th,
VALUE name);
757#define native_set_another_thread_name rb_w32_set_thread_description_str
763#define TIMER_THREAD_CREATED_P() (timer_thread.id != 0)
765static unsigned long __stdcall
766timer_thread_func(
void *dummy)
769 RUBY_DEBUG_LOG(
"start");
770 rb_w32_set_thread_description(GetCurrentThread(), L
"ruby-timer-thread");
771 while (WaitForSingleObject(timer_thread.lock,
772 TIME_QUANTUM_USEC/1000) == WAIT_TIMEOUT) {
774 rb_threadptr_check_signal(vm->ractor.main_thread);
776 RUBY_DEBUG_LOG(
"end");
781rb_thread_wakeup_timer_thread(
int sig)
787rb_thread_create_timer_thread(
void)
789 if (timer_thread.id == 0) {
790 if (!timer_thread.lock) {
791 timer_thread.lock = CreateEvent(0, TRUE, FALSE, 0);
793 timer_thread.id = w32_create_thread(1024 + (USE_RUBY_DEBUG_LOG ? BUFSIZ : 0),
794 timer_thread_func, 0);
795 w32_resume_thread(timer_thread.id);
800native_stop_timer_thread(
void)
802 int stopped = --system_working <= 0;
804 SetEvent(timer_thread.lock);
805 native_thread_join(timer_thread.id);
806 CloseHandle(timer_thread.lock);
807 timer_thread.lock = 0;
813native_reset_timer_thread(
void)
815 if (timer_thread.id) {
816 CloseHandle(timer_thread.id);
822ruby_stack_overflowed_p(
const rb_thread_t *th,
const void *addr)
824 return rb_ec_raised_p(th->ec, RAISED_STACKOVERFLOW);
827#if defined(__MINGW32__)
829rb_w32_stack_overflow_handler(
struct _EXCEPTION_POINTERS *exception)
831 if (exception->ExceptionRecord->ExceptionCode == EXCEPTION_STACK_OVERFLOW) {
832 rb_ec_raised_set(GET_EC(), RAISED_STACKOVERFLOW);
835 return EXCEPTION_CONTINUE_SEARCH;
839#ifdef RUBY_ALLOCA_CHKSTK
841ruby_alloca_chkstk(
size_t len,
void *sp)
845 if (!rb_ec_raised_p(ec, RAISED_STACKOVERFLOW)) {
846 rb_ec_raised_set(ec, RAISED_STACKOVERFLOW);
847 rb_exc_raise(sysstack_error);
861 return GetCurrentThread();
872 DWORD tid = GetThreadId(th->nt->thread_id);
873 if (tid == 0) rb_sys_fail(
"GetThreadId");
876#define USE_NATIVE_THREAD_NATIVE_THREAD_ID 1
898 native_thread_destroy(th->nt);
900 ruby_xfree(th->sched.vm_stack);
910rb_thread_sched_mark_zombies(
rb_vm_t *vm)
916vm_barrier_finish_p(
rb_vm_t *vm)
918 RUBY_DEBUG_LOG(
"cnt:%u living:%u blocking:%u",
919 vm->ractor.blocking_cnt == vm->ractor.cnt,
920 vm->ractor.sync.barrier_cnt,
922 vm->ractor.blocking_cnt);
924 VM_ASSERT(vm->ractor.blocking_cnt <= vm->ractor.cnt);
925 return vm->ractor.blocking_cnt == vm->ractor.cnt;
931 vm->ractor.sync.barrier_waiting =
true;
933 RUBY_DEBUG_LOG(
"barrier start. cnt:%u living:%u blocking:%u",
934 vm->ractor.sync.barrier_cnt,
936 vm->ractor.blocking_cnt);
938 rb_vm_ractor_blocking_cnt_inc(vm, cr, __FILE__, __LINE__);
942 ccan_list_for_each(&vm->ractor.set, r, vmlr_node) {
944 rb_ractor_vm_barrier_interrupt_running_thread(r);
949 while (!vm_barrier_finish_p(vm)) {
950 rb_vm_cond_wait(vm, &vm->ractor.sync.barrier_cond);
953 RUBY_DEBUG_LOG(
"cnt:%u barrier success", vm->ractor.sync.barrier_cnt);
955 rb_vm_ractor_blocking_cnt_dec(vm, cr, __FILE__, __LINE__);
957 vm->ractor.sync.barrier_waiting =
false;
958 vm->ractor.sync.barrier_cnt++;
960 ccan_list_for_each(&vm->ractor.set, r, vmlr_node) {
968 vm->ractor.sync.lock_owner = cr;
969 unsigned int barrier_cnt = vm->ractor.sync.barrier_cnt;
973 RB_VM_SAVE_MACHINE_CONTEXT(th);
975 if (rb_ractor_status_p(cr, ractor_running)) {
976 rb_vm_ractor_blocking_cnt_inc(vm, cr, __FILE__, __LINE__);
982 VM_ASSERT(rb_ractor_status_p(cr, ractor_blocking));
984 if (vm_barrier_finish_p(vm)) {
985 RUBY_DEBUG_LOG(
"wakeup barrier owner");
989 RUBY_DEBUG_LOG(
"wait for barrier finish");
993 while (barrier_cnt == vm->ractor.sync.barrier_cnt) {
994 vm->ractor.sync.lock_owner = NULL;
996 VM_ASSERT(vm->ractor.sync.lock_owner == NULL);
997 vm->ractor.sync.lock_owner = cr;
1000 RUBY_DEBUG_LOG(
"barrier is released. Acquire vm_lock");
1003 rb_vm_ractor_blocking_cnt_dec(vm, cr, __FILE__, __LINE__);
1006 vm->ractor.sync.lock_owner = NULL;
uint32_t rb_event_flag_t
Represents event(s).
#define UNREACHABLE
Old name of RBIMPL_UNREACHABLE.
#define ULONG2NUM
Old name of RB_ULONG2NUM.
#define ZALLOC
Old name of RB_ZALLOC.
void ruby_init_stack(volatile VALUE *addr)
Set stack bottom of Ruby implementation.
size_t ruby_stack_length(VALUE **p)
Queries what Ruby thinks is the machine stack.
int rb_reserved_fd_p(int fd)
Queries if the given FD is reserved or not.
int len
Length of the buffer.
rb_internal_thread_event_hook_t * rb_internal_thread_add_event_hook(rb_internal_thread_event_callback func, rb_event_flag_t events, void *data)
Registers a thread event hook function.
bool rb_internal_thread_remove_event_hook(rb_internal_thread_event_hook_t *hook)
Unregister the passed hook.
static fd_set * rb_fd_ptr(const rb_fdset_t *f)
Raw pointer to fd_set.
#define ALLOCA_N(type, n)
#define RBIMPL_ATTR_NORETURN()
Wraps (or simulates) [[noreturn]]
#define errno
Ractor-aware version of errno.
#define rb_fd_resize(n, f)
Does nothing (defined for compatibility).
The data structure which wraps the fd_set bitmap used by select(2).
rb_nativethread_id_t rb_nativethread_self(void)
Queries the ID of the native thread that is calling this function.
void rb_native_mutex_lock(rb_nativethread_lock_t *lock)
Just another name of rb_nativethread_lock_lock.
void rb_native_cond_initialize(rb_nativethread_cond_t *cond)
Fills the passed condition variable with an initial value.
int rb_native_mutex_trylock(rb_nativethread_lock_t *lock)
Identical to rb_native_mutex_lock(), except it doesn't block in case rb_native_mutex_lock() would.
void rb_native_cond_broadcast(rb_nativethread_cond_t *cond)
Signals a condition variable.
void rb_native_mutex_initialize(rb_nativethread_lock_t *lock)
Just another name of rb_nativethread_lock_initialize.
void rb_native_mutex_unlock(rb_nativethread_lock_t *lock)
Just another name of rb_nativethread_lock_unlock.
void rb_native_mutex_destroy(rb_nativethread_lock_t *lock)
Just another name of rb_nativethread_lock_destroy.
void rb_native_cond_destroy(rb_nativethread_cond_t *cond)
Destroys the passed condition variable.
void rb_native_cond_signal(rb_nativethread_cond_t *cond)
Signals a condition variable.
void rb_native_cond_wait(rb_nativethread_cond_t *cond, rb_nativethread_lock_t *mutex)
Waits for the passed condition variable to be signalled.
void rb_native_cond_timedwait(rb_nativethread_cond_t *cond, rb_nativethread_lock_t *mutex, unsigned long msec)
Identical to rb_native_cond_wait(), except it additionally takes timeout in msec resolution.
uintptr_t VALUE
Type that represents a Ruby object.