| /* Copyright Joyent, Inc. and other Node contributors. All rights reserved. |
| * |
| * Permission is hereby granted, free of charge, to any person obtaining a copy |
| * of this software and associated documentation files (the "Software"), to |
| * deal in the Software without restriction, including without limitation the |
| * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or |
| * sell copies of the Software, and to permit persons to whom the Software is |
| * furnished to do so, subject to the following conditions: |
| * |
| * The above copyright notice and this permission notice shall be included in |
| * all copies or substantial portions of the Software. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
| * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE |
| * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
| * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING |
| * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS |
| * IN THE SOFTWARE. |
| */ |
| |
| #include "uv.h" |
| #include "internal.h" |
| |
| #include <pthread.h> |
| #include <assert.h> |
| #include <errno.h> |
| |
| #if defined(__APPLE__) && defined(__MACH__) |
| #include <sys/time.h> |
| #endif /* defined(__APPLE__) && defined(__MACH__) */ |
| |
| #undef NANOSEC |
| #define NANOSEC ((uint64_t) 1e9) |
| |
| int uv_thread_join(uv_thread_t *tid) { |
| if (pthread_join(*tid, NULL)) |
| return -1; |
| else |
| return 0; |
| } |
| |
| |
| int uv_mutex_init(uv_mutex_t* mutex) { |
| #if defined(NDEBUG) || !defined(PTHREAD_MUTEX_ERRORCHECK) |
| if (pthread_mutex_init(mutex, NULL)) |
| return -1; |
| else |
| return 0; |
| #else |
| pthread_mutexattr_t attr; |
| int r; |
| |
| if (pthread_mutexattr_init(&attr)) |
| abort(); |
| |
| if (pthread_mutexattr_settype(&attr, PTHREAD_MUTEX_ERRORCHECK)) |
| abort(); |
| |
| r = pthread_mutex_init(mutex, &attr); |
| |
| if (pthread_mutexattr_destroy(&attr)) |
| abort(); |
| |
| return r ? -1 : 0; |
| #endif |
| } |
| |
| |
| void uv_mutex_destroy(uv_mutex_t* mutex) { |
| if (pthread_mutex_destroy(mutex)) |
| abort(); |
| } |
| |
| |
| void uv_mutex_lock(uv_mutex_t* mutex) { |
| if (pthread_mutex_lock(mutex)) |
| abort(); |
| } |
| |
| |
| int uv_mutex_trylock(uv_mutex_t* mutex) { |
| int r; |
| |
| r = pthread_mutex_trylock(mutex); |
| |
| if (r && r != EBUSY && r != EAGAIN) |
| abort(); |
| |
| if (r) |
| return -1; |
| else |
| return 0; |
| } |
| |
| |
| void uv_mutex_unlock(uv_mutex_t* mutex) { |
| if (pthread_mutex_unlock(mutex)) |
| abort(); |
| } |
| |
| |
| int uv_rwlock_init(uv_rwlock_t* rwlock) { |
| if (pthread_rwlock_init(rwlock, NULL)) |
| return -1; |
| else |
| return 0; |
| } |
| |
| |
| void uv_rwlock_destroy(uv_rwlock_t* rwlock) { |
| if (pthread_rwlock_destroy(rwlock)) |
| abort(); |
| } |
| |
| |
| void uv_rwlock_rdlock(uv_rwlock_t* rwlock) { |
| if (pthread_rwlock_rdlock(rwlock)) |
| abort(); |
| } |
| |
| |
| int uv_rwlock_tryrdlock(uv_rwlock_t* rwlock) { |
| int r; |
| |
| r = pthread_rwlock_tryrdlock(rwlock); |
| |
| if (r && r != EBUSY && r != EAGAIN) |
| abort(); |
| |
| if (r) |
| return -1; |
| else |
| return 0; |
| } |
| |
| |
| void uv_rwlock_rdunlock(uv_rwlock_t* rwlock) { |
| if (pthread_rwlock_unlock(rwlock)) |
| abort(); |
| } |
| |
| |
| void uv_rwlock_wrlock(uv_rwlock_t* rwlock) { |
| if (pthread_rwlock_wrlock(rwlock)) |
| abort(); |
| } |
| |
| |
| int uv_rwlock_trywrlock(uv_rwlock_t* rwlock) { |
| int r; |
| |
| r = pthread_rwlock_trywrlock(rwlock); |
| |
| if (r && r != EBUSY && r != EAGAIN) |
| abort(); |
| |
| if (r) |
| return -1; |
| else |
| return 0; |
| } |
| |
| |
| void uv_rwlock_wrunlock(uv_rwlock_t* rwlock) { |
| if (pthread_rwlock_unlock(rwlock)) |
| abort(); |
| } |
| |
| |
| void uv_once(uv_once_t* guard, void (*callback)(void)) { |
| if (pthread_once(guard, callback)) |
| abort(); |
| } |
| |
| #if defined(__APPLE__) && defined(__MACH__) |
| |
| int uv_sem_init(uv_sem_t* sem, unsigned int value) { |
| if (semaphore_create(mach_task_self(), sem, SYNC_POLICY_FIFO, value)) |
| return -1; |
| else |
| return 0; |
| } |
| |
| |
| void uv_sem_destroy(uv_sem_t* sem) { |
| if (semaphore_destroy(mach_task_self(), *sem)) |
| abort(); |
| } |
| |
| |
| void uv_sem_post(uv_sem_t* sem) { |
| if (semaphore_signal(*sem)) |
| abort(); |
| } |
| |
| |
| void uv_sem_wait(uv_sem_t* sem) { |
| int r; |
| |
| do |
| r = semaphore_wait(*sem); |
| while (r == KERN_ABORTED); |
| |
| if (r != KERN_SUCCESS) |
| abort(); |
| } |
| |
| |
| int uv_sem_trywait(uv_sem_t* sem) { |
| mach_timespec_t interval; |
| |
| interval.tv_sec = 0; |
| interval.tv_nsec = 0; |
| |
| if (semaphore_timedwait(*sem, interval) == KERN_SUCCESS) |
| return 0; |
| else |
| return -1; |
| } |
| |
| #else /* !(defined(__APPLE__) && defined(__MACH__)) */ |
| |
| int uv_sem_init(uv_sem_t* sem, unsigned int value) { |
| return sem_init(sem, 0, value); |
| } |
| |
| |
| void uv_sem_destroy(uv_sem_t* sem) { |
| if (sem_destroy(sem)) |
| abort(); |
| } |
| |
| |
| void uv_sem_post(uv_sem_t* sem) { |
| if (sem_post(sem)) |
| abort(); |
| } |
| |
| |
| void uv_sem_wait(uv_sem_t* sem) { |
| int r; |
| |
| do |
| r = sem_wait(sem); |
| while (r == -1 && errno == EINTR); |
| |
| if (r) |
| abort(); |
| } |
| |
| |
| int uv_sem_trywait(uv_sem_t* sem) { |
| int r; |
| |
| do |
| r = sem_trywait(sem); |
| while (r == -1 && errno == EINTR); |
| |
| if (r && errno != EAGAIN) |
| abort(); |
| |
| return r; |
| } |
| |
| #endif /* defined(__APPLE__) && defined(__MACH__) */ |
| |
| |
| #if defined(__APPLE__) && defined(__MACH__) |
| |
| int uv_cond_init(uv_cond_t* cond) { |
| if (pthread_cond_init(cond, NULL)) |
| return -1; |
| else |
| return 0; |
| } |
| |
| #else /* !(defined(__APPLE__) && defined(__MACH__)) */ |
| |
| int uv_cond_init(uv_cond_t* cond) { |
| pthread_condattr_t attr; |
| |
| if (pthread_condattr_init(&attr)) |
| return -1; |
| |
| if (pthread_condattr_setclock(&attr, CLOCK_MONOTONIC)) |
| goto error2; |
| |
| if (pthread_cond_init(cond, &attr)) |
| goto error2; |
| |
| if (pthread_condattr_destroy(&attr)) |
| goto error; |
| |
| return 0; |
| |
| error: |
| pthread_cond_destroy(cond); |
| error2: |
| pthread_condattr_destroy(&attr); |
| return -1; |
| } |
| |
| #endif /* defined(__APPLE__) && defined(__MACH__) */ |
| |
| void uv_cond_destroy(uv_cond_t* cond) { |
| if (pthread_cond_destroy(cond)) |
| abort(); |
| } |
| |
| void uv_cond_signal(uv_cond_t* cond) { |
| if (pthread_cond_signal(cond)) |
| abort(); |
| } |
| |
| void uv_cond_broadcast(uv_cond_t* cond) { |
| if (pthread_cond_broadcast(cond)) |
| abort(); |
| } |
| |
| void uv_cond_wait(uv_cond_t* cond, uv_mutex_t* mutex) { |
| if (pthread_cond_wait(cond, mutex)) |
| abort(); |
| } |
| |
| |
| int uv_cond_timedwait(uv_cond_t* cond, uv_mutex_t* mutex, uint64_t timeout) { |
| int r; |
| struct timespec ts; |
| |
| #if defined(__APPLE__) && defined(__MACH__) |
| ts.tv_sec = timeout / NANOSEC; |
| ts.tv_nsec = timeout % NANOSEC; |
| r = pthread_cond_timedwait_relative_np(cond, mutex, &ts); |
| #else |
| timeout += uv__hrtime(); |
| ts.tv_sec = timeout / NANOSEC; |
| ts.tv_nsec = timeout % NANOSEC; |
| r = pthread_cond_timedwait(cond, mutex, &ts); |
| #endif |
| |
| |
| if (r == 0) |
| return 0; |
| |
| if (r == ETIMEDOUT) |
| return -1; |
| |
| abort(); |
| return -1; /* Satisfy the compiler. */ |
| } |
| |
| |
| #if defined(__APPLE__) && defined(__MACH__) |
| |
| int uv_barrier_init(uv_barrier_t* barrier, unsigned int count) { |
| barrier->n = count; |
| barrier->count = 0; |
| |
| if (uv_mutex_init(&barrier->mutex)) |
| return -1; |
| |
| if (uv_sem_init(&barrier->turnstile1, 0)) |
| goto error2; |
| |
| if (uv_sem_init(&barrier->turnstile2, 1)) |
| goto error; |
| |
| return 0; |
| |
| error: |
| uv_sem_destroy(&barrier->turnstile1); |
| error2: |
| uv_mutex_destroy(&barrier->mutex); |
| return -1; |
| |
| } |
| |
| |
| void uv_barrier_destroy(uv_barrier_t* barrier) { |
| uv_sem_destroy(&barrier->turnstile2); |
| uv_sem_destroy(&barrier->turnstile1); |
| uv_mutex_destroy(&barrier->mutex); |
| } |
| |
| |
| void uv_barrier_wait(uv_barrier_t* barrier) { |
| uv_mutex_lock(&barrier->mutex); |
| if (++barrier->count == barrier->n) { |
| uv_sem_wait(&barrier->turnstile2); |
| uv_sem_post(&barrier->turnstile1); |
| } |
| uv_mutex_unlock(&barrier->mutex); |
| |
| uv_sem_wait(&barrier->turnstile1); |
| uv_sem_post(&barrier->turnstile1); |
| |
| uv_mutex_lock(&barrier->mutex); |
| if (--barrier->count == 0) { |
| uv_sem_wait(&barrier->turnstile1); |
| uv_sem_post(&barrier->turnstile2); |
| } |
| uv_mutex_unlock(&barrier->mutex); |
| |
| uv_sem_wait(&barrier->turnstile2); |
| uv_sem_post(&barrier->turnstile2); |
| } |
| |
| #else /* !(defined(__APPLE__) && defined(__MACH__)) */ |
| |
| int uv_barrier_init(uv_barrier_t* barrier, unsigned int count) { |
| if (pthread_barrier_init(barrier, NULL, count)) |
| return -1; |
| else |
| return 0; |
| } |
| |
| |
| void uv_barrier_destroy(uv_barrier_t* barrier) { |
| if (pthread_barrier_destroy(barrier)) |
| abort(); |
| } |
| |
| |
| void uv_barrier_wait(uv_barrier_t* barrier) { |
| int r = pthread_barrier_wait(barrier); |
| if (r && r != PTHREAD_BARRIER_SERIAL_THREAD) |
| abort(); |
| } |
| |
| #endif /* defined(__APPLE__) && defined(__MACH__) */ |