#include "lint.h"
#include "thr_uberdata.h"
#include <sys/sdt.h>
#define TRY_FLAG 0x10
#define READ_LOCK 0
#define WRITE_LOCK 1
#define READ_LOCK_TRY (READ_LOCK | TRY_FLAG)
#define WRITE_LOCK_TRY (WRITE_LOCK | TRY_FLAG)
#define NLOCKS 4
#define ASSERT_CONSISTENT_STATE(readers) \
ASSERT(!((readers) & URW_WRITE_LOCKED) || \
((readers) & ~URW_HAS_WAITERS) == URW_WRITE_LOCKED)
static readlock_t *
rwl_entry(rwlock_t *rwlp)
{
ulwp_t *self = curthread;
readlock_t *remembered = NULL;
readlock_t *readlockp;
uint_t nlocks;
ASSERT((self->ul_critical + self->ul_sigdefer) != 0);
if ((nlocks = self->ul_rdlockcnt) != 0)
readlockp = self->ul_readlock.array;
else {
nlocks = 1;
readlockp = &self->ul_readlock.single;
}
for (; nlocks; nlocks--, readlockp++) {
if (readlockp->rd_rwlock == rwlp)
return (readlockp);
if (readlockp->rd_count == 0 && remembered == NULL)
remembered = readlockp;
}
if (remembered != NULL) {
remembered->rd_rwlock = rwlp;
return (remembered);
}
if ((nlocks = self->ul_rdlockcnt) == 0) {
self->ul_rdlockcnt = nlocks = NLOCKS;
readlockp = lmalloc(nlocks * sizeof (readlock_t));
*readlockp = self->ul_readlock.single;
self->ul_readlock.single.rd_count = 0;
self->ul_readlock.array = readlockp;
(++readlockp)->rd_rwlock = rwlp;
return (readlockp);
}
readlockp = lmalloc(nlocks * 2 * sizeof (readlock_t));
(void) memcpy(readlockp, self->ul_readlock.array,
nlocks * sizeof (readlock_t));
lfree(self->ul_readlock.array, nlocks * sizeof (readlock_t));
self->ul_readlock.array = readlockp;
self->ul_rdlockcnt *= 2;
(readlockp += nlocks)->rd_rwlock = rwlp;
return (readlockp);
}
void
rwl_free(ulwp_t *ulwp)
{
uint_t nlocks;
if ((nlocks = ulwp->ul_rdlockcnt) != 0)
lfree(ulwp->ul_readlock.array, nlocks * sizeof (readlock_t));
ulwp->ul_rdlockcnt = 0;
ulwp->ul_readlock.single.rd_rwlock = NULL;
ulwp->ul_readlock.single.rd_count = 0;
}
#pragma weak _rw_read_held = rw_read_held
int
rw_read_held(rwlock_t *rwlp)
{
volatile uint32_t *rwstate = (volatile uint32_t *)&rwlp->rwlock_readers;
uint32_t readers;
ulwp_t *self = curthread;
readlock_t *readlockp;
uint_t nlocks;
int rval = 0;
no_preempt(self);
readers = *rwstate;
ASSERT_CONSISTENT_STATE(readers);
if (!(readers & URW_WRITE_LOCKED) &&
(readers & URW_READERS_MASK) != 0) {
if ((nlocks = self->ul_rdlockcnt) != 0)
readlockp = self->ul_readlock.array;
else {
nlocks = 1;
readlockp = &self->ul_readlock.single;
}
for (; nlocks; nlocks--, readlockp++) {
if (readlockp->rd_rwlock == rwlp) {
if (readlockp->rd_count)
rval = 1;
break;
}
}
}
preempt(self);
return (rval);
}
#pragma weak _rw_write_held = rw_write_held
int
rw_write_held(rwlock_t *rwlp)
{
volatile uint32_t *rwstate = (volatile uint32_t *)&rwlp->rwlock_readers;
uint32_t readers;
ulwp_t *self = curthread;
int rval;
no_preempt(self);
readers = *rwstate;
ASSERT_CONSISTENT_STATE(readers);
rval = ((readers & URW_WRITE_LOCKED) &&
rwlp->rwlock_owner == (uintptr_t)self &&
(rwlp->rwlock_type == USYNC_THREAD ||
rwlp->rwlock_ownerpid == self->ul_uberdata->pid));
preempt(self);
return (rval);
}
#pragma weak _rwlock_init = rwlock_init
int
rwlock_init(rwlock_t *rwlp, int type, void *arg __unused)
{
ulwp_t *self = curthread;
if (type != USYNC_THREAD && type != USYNC_PROCESS)
return (EINVAL);
sigoff(self);
rwl_entry(rwlp)->rd_count = 0;
sigon(self);
(void) memset(rwlp, 0, sizeof (*rwlp));
rwlp->rwlock_type = (uint16_t)type;
rwlp->rwlock_magic = RWL_MAGIC;
rwlp->mutex.mutex_type = (uint8_t)type;
rwlp->mutex.mutex_flag = LOCK_INITED;
rwlp->mutex.mutex_magic = MUTEX_MAGIC;
if (((uintptr_t)rwlp & (_LONG_LONG_ALIGNMENT - 1)) &&
self->ul_misaligned == 0)
return (EINVAL);
return (0);
}
#pragma weak pthread_rwlock_destroy = rwlock_destroy
#pragma weak _rwlock_destroy = rwlock_destroy
int
rwlock_destroy(rwlock_t *rwlp)
{
ulwp_t *self = curthread;
sigoff(self);
rwl_entry(rwlp)->rd_count = 0;
sigon(self);
rwlp->rwlock_magic = 0;
tdb_sync_obj_deregister(rwlp);
return (0);
}
static int
read_lock_try(rwlock_t *rwlp, int ignore_waiters_flag)
{
volatile uint32_t *rwstate = (volatile uint32_t *)&rwlp->rwlock_readers;
uint32_t mask = ignore_waiters_flag?
URW_WRITE_LOCKED : (URW_HAS_WAITERS | URW_WRITE_LOCKED);
uint32_t readers;
ulwp_t *self = curthread;
no_preempt(self);
while (((readers = *rwstate) & mask) == 0) {
if (atomic_cas_32(rwstate, readers, readers + 1) == readers) {
preempt(self);
return (1);
}
}
preempt(self);
return (0);
}
static int
read_unlock_try(rwlock_t *rwlp)
{
volatile uint32_t *rwstate = (volatile uint32_t *)&rwlp->rwlock_readers;
uint32_t readers;
ulwp_t *self = curthread;
no_preempt(self);
while (((readers = *rwstate) & URW_HAS_WAITERS) == 0) {
if (atomic_cas_32(rwstate, readers, readers - 1) == readers) {
preempt(self);
return (1);
}
}
preempt(self);
return (0);
}
static int
write_lock_try(rwlock_t *rwlp, int ignore_waiters_flag)
{
volatile uint32_t *rwstate = (volatile uint32_t *)&rwlp->rwlock_readers;
uint32_t mask = ignore_waiters_flag?
(URW_WRITE_LOCKED | URW_READERS_MASK) :
(URW_HAS_WAITERS | URW_WRITE_LOCKED | URW_READERS_MASK);
ulwp_t *self = curthread;
uint32_t readers;
no_preempt(self);
while (((readers = *rwstate) & mask) == 0) {
if (atomic_cas_32(rwstate, readers, readers | URW_WRITE_LOCKED)
== readers) {
preempt(self);
return (1);
}
}
preempt(self);
return (0);
}
static int
write_unlock_try(rwlock_t *rwlp)
{
volatile uint32_t *rwstate = (volatile uint32_t *)&rwlp->rwlock_readers;
uint32_t readers;
ulwp_t *self = curthread;
no_preempt(self);
while (((readers = *rwstate) & URW_HAS_WAITERS) == 0) {
if (atomic_cas_32(rwstate, readers, 0) == readers) {
preempt(self);
return (1);
}
}
preempt(self);
return (0);
}
static void
rw_queue_release(rwlock_t *rwlp)
{
volatile uint32_t *rwstate = (volatile uint32_t *)&rwlp->rwlock_readers;
queue_head_t *qp;
uint32_t readers;
uint32_t writer;
ulwp_t **ulwpp;
ulwp_t *ulwp;
ulwp_t *prev;
int nlwpid = 0;
int more;
int maxlwps = MAXLWPS;
lwpid_t buffer[MAXLWPS];
lwpid_t *lwpid = buffer;
qp = queue_lock(rwlp, MX);
readers = *rwstate;
ASSERT_CONSISTENT_STATE(readers);
if (readers & URW_WRITE_LOCKED)
atomic_and_32(rwstate, ~URW_WRITE_LOCKED);
else
atomic_dec_32(rwstate);
if (!(readers & URW_HAS_WAITERS)) {
queue_unlock(qp);
return;
}
readers = *rwstate;
ASSERT_CONSISTENT_STATE(readers);
ASSERT(readers & URW_HAS_WAITERS);
readers &= URW_READERS_MASK;
writer = 0;
while ((ulwpp = queue_slot(qp, &prev, &more)) != NULL) {
ulwp = *ulwpp;
ASSERT(ulwp->ul_wchan == rwlp);
if (ulwp->ul_writer) {
if (writer != 0 || readers != 0)
break;
writer++;
} else {
if (writer != 0)
break;
readers++;
if (nlwpid == maxlwps)
lwpid = alloc_lwpids(lwpid, &nlwpid, &maxlwps);
}
queue_unlink(qp, ulwpp, prev);
ulwp->ul_sleepq = NULL;
ulwp->ul_wchan = NULL;
if (writer) {
ASSERT((*rwstate & ~URW_HAS_WAITERS) == 0);
atomic_or_32(rwstate, URW_WRITE_LOCKED);
rwlp->rwlock_owner = (uintptr_t)ulwp;
}
lwpid[nlwpid++] = ulwp->ul_lwpid;
}
if (ulwpp == NULL)
atomic_and_32(rwstate, ~URW_HAS_WAITERS);
if (nlwpid == 0) {
queue_unlock(qp);
} else {
ulwp_t *self = curthread;
no_preempt(self);
queue_unlock(qp);
if (nlwpid == 1)
(void) __lwp_unpark(lwpid[0]);
else
(void) __lwp_unpark_all(lwpid, nlwpid);
preempt(self);
}
if (lwpid != buffer)
(void) munmap((caddr_t)lwpid, maxlwps * sizeof (lwpid_t));
}
int
shared_rwlock_lock(rwlock_t *rwlp, timespec_t *tsp, int rd_wr)
{
volatile uint32_t *rwstate = (volatile uint32_t *)&rwlp->rwlock_readers;
mutex_t *mp = &rwlp->mutex;
int try_flag;
int error;
try_flag = (rd_wr & TRY_FLAG);
rd_wr &= ~TRY_FLAG;
ASSERT(rd_wr == READ_LOCK || rd_wr == WRITE_LOCK);
if (!try_flag) {
DTRACE_PROBE2(plockstat, rw__block, rwlp, rd_wr);
}
do {
if (try_flag && (*rwstate & URW_WRITE_LOCKED)) {
error = EBUSY;
break;
}
if ((error = mutex_lock(mp)) != 0)
break;
if (rd_wr == READ_LOCK) {
if (read_lock_try(rwlp, 0)) {
(void) mutex_unlock(mp);
break;
}
} else {
if (write_lock_try(rwlp, 0)) {
(void) mutex_unlock(mp);
break;
}
}
atomic_or_32(rwstate, URW_HAS_WAITERS);
#ifdef DEBUG
uint32_t readers;
readers = *rwstate;
ASSERT_CONSISTENT_STATE(readers);
#endif
DTRACE_PROBE2(plockstat, mutex__release, mp, 0);
if (rd_wr == READ_LOCK) {
if (try_flag)
error = __lwp_rwlock_tryrdlock(rwlp);
else
error = __lwp_rwlock_rdlock(rwlp, tsp);
} else {
if (try_flag)
error = __lwp_rwlock_trywrlock(rwlp);
else
error = __lwp_rwlock_wrlock(rwlp, tsp);
}
} while (error == EAGAIN || error == EINTR);
if (!try_flag) {
DTRACE_PROBE3(plockstat, rw__blocked, rwlp, rd_wr, error == 0);
}
return (error);
}
int
rwlock_lock(rwlock_t *rwlp, timespec_t *tsp, int rd_wr)
{
volatile uint32_t *rwstate = (volatile uint32_t *)&rwlp->rwlock_readers;
uint32_t readers;
ulwp_t *self = curthread;
queue_head_t *qp;
ulwp_t *ulwp;
int try_flag;
int ignore_waiters_flag;
int error = 0;
try_flag = (rd_wr & TRY_FLAG);
rd_wr &= ~TRY_FLAG;
ASSERT(rd_wr == READ_LOCK || rd_wr == WRITE_LOCK);
if (!try_flag) {
DTRACE_PROBE2(plockstat, rw__block, rwlp, rd_wr);
}
qp = queue_lock(rwlp, MX);
ignore_waiters_flag = 0;
while (error == 0) {
if (rd_wr == READ_LOCK) {
if (read_lock_try(rwlp, ignore_waiters_flag))
break;
} else {
if (write_lock_try(rwlp, ignore_waiters_flag))
break;
}
ignore_waiters_flag = 1;
atomic_or_32(rwstate, URW_HAS_WAITERS);
readers = *rwstate;
ASSERT_CONSISTENT_STATE(readers);
if ((readers & URW_WRITE_LOCKED) ||
(rd_wr == WRITE_LOCK &&
(readers & URW_READERS_MASK) != 0))
;
else if ((ulwp = queue_waiter(qp)) == NULL) {
atomic_and_32(rwstate, ~URW_HAS_WAITERS);
ignore_waiters_flag = 0;
continue;
} else {
int our_pri = real_priority(self);
int his_pri = real_priority(ulwp);
if (rd_wr == WRITE_LOCK) {
if (his_pri <= our_pri) {
continue;
}
} else {
if (his_pri < our_pri ||
(his_pri == our_pri && !ulwp->ul_writer)) {
continue;
}
}
}
if (try_flag) {
error = EBUSY;
break;
}
self->ul_writer = rd_wr;
enqueue(qp, self, 0);
set_parking_flag(self, 1);
queue_unlock(qp);
if ((error = __lwp_park(tsp, 0)) == EINTR)
error = 0;
set_parking_flag(self, 0);
qp = queue_lock(rwlp, MX);
if (self->ul_sleepq && dequeue_self(qp) == 0) {
atomic_and_32(rwstate, ~URW_HAS_WAITERS);
ignore_waiters_flag = 0;
}
self->ul_writer = 0;
if (rd_wr == WRITE_LOCK &&
(*rwstate & URW_WRITE_LOCKED) &&
rwlp->rwlock_owner == (uintptr_t)self) {
error = 0;
break;
}
}
if (qp->qh_root == NULL || qp->qh_root->qr_head == NULL)
atomic_and_32(rwstate, ~URW_HAS_WAITERS);
queue_unlock(qp);
if (!try_flag) {
DTRACE_PROBE3(plockstat, rw__blocked, rwlp, rd_wr, error == 0);
}
return (error);
}
int
rw_rdlock_impl(rwlock_t *rwlp, timespec_t *tsp)
{
ulwp_t *self = curthread;
uberdata_t *udp = self->ul_uberdata;
readlock_t *readlockp;
tdb_rwlock_stats_t *rwsp = RWLOCK_STATS(rwlp, udp);
int error;
sigoff(self);
readlockp = rwl_entry(rwlp);
if (readlockp->rd_count != 0) {
if (readlockp->rd_count == READ_LOCK_MAX) {
sigon(self);
error = EAGAIN;
goto out;
}
sigon(self);
error = 0;
goto out;
}
sigon(self);
if (rw_write_held(rwlp)) {
if (self->ul_error_detection)
rwlock_error(rwlp, "rwlock_rdlock",
"calling thread owns the writer lock");
error = EDEADLK;
goto out;
}
if (read_lock_try(rwlp, 0))
error = 0;
else if (rwlp->rwlock_type == USYNC_PROCESS)
error = shared_rwlock_lock(rwlp, tsp, READ_LOCK);
else
error = rwlock_lock(rwlp, tsp, READ_LOCK);
out:
if (error == 0) {
sigoff(self);
rwl_entry(rwlp)->rd_count++;
sigon(self);
if (rwsp)
tdb_incr(rwsp->rw_rdlock);
DTRACE_PROBE2(plockstat, rw__acquire, rwlp, READ_LOCK);
} else {
DTRACE_PROBE3(plockstat, rw__error, rwlp, READ_LOCK, error);
}
return (error);
}
#pragma weak pthread_rwlock_rdlock = rw_rdlock
#pragma weak _rw_rdlock = rw_rdlock
int
rw_rdlock(rwlock_t *rwlp)
{
ASSERT(!curthread->ul_critical || curthread->ul_bindflags);
return (rw_rdlock_impl(rwlp, NULL));
}
void
lrw_rdlock(rwlock_t *rwlp)
{
enter_critical(curthread);
(void) rw_rdlock_impl(rwlp, NULL);
}
int
pthread_rwlock_relclockrdlock_np(pthread_rwlock_t *restrict rwlp,
clockid_t clock, const struct timespec *restrict reltime)
{
timespec_t tslocal = *reltime;
int error;
ASSERT(!curthread->ul_critical || curthread->ul_bindflags);
switch (clock) {
case CLOCK_REALTIME:
case CLOCK_HIGHRES:
break;
default:
return (EINVAL);
}
error = rw_rdlock_impl((rwlock_t *)rwlp, &tslocal);
if (error == ETIME)
error = ETIMEDOUT;
return (error);
}
int
pthread_rwlock_reltimedrdlock_np(pthread_rwlock_t *restrict rwlp,
const struct timespec *restrict reltime)
{
return (pthread_rwlock_relclockrdlock_np(rwlp, CLOCK_REALTIME,
reltime));
}
int
pthread_rwlock_clockrdlock(pthread_rwlock_t *restrict rwlp, clockid_t clock,
const struct timespec *restrict abstime)
{
timespec_t tslocal;
int error;
ASSERT(!curthread->ul_critical || curthread->ul_bindflags);
switch (clock) {
case CLOCK_REALTIME:
case CLOCK_HIGHRES:
break;
default:
return (EINVAL);
}
abstime_to_reltime(clock, abstime, &tslocal);
error = rw_rdlock_impl((rwlock_t *)rwlp, &tslocal);
if (error == ETIME)
error = ETIMEDOUT;
return (error);
}
int
pthread_rwlock_timedrdlock(pthread_rwlock_t *restrict rwlp,
const struct timespec *restrict abstime)
{
return (pthread_rwlock_clockrdlock(rwlp, CLOCK_REALTIME, abstime));
}
int
rw_wrlock_impl(rwlock_t *rwlp, timespec_t *tsp)
{
ulwp_t *self = curthread;
uberdata_t *udp = self->ul_uberdata;
tdb_rwlock_stats_t *rwsp = RWLOCK_STATS(rwlp, udp);
int error;
if (rw_read_held(rwlp)) {
if (self->ul_error_detection)
rwlock_error(rwlp, "rwlock_wrlock",
"calling thread owns the readers lock");
error = EDEADLK;
goto out;
}
if (rw_write_held(rwlp)) {
if (self->ul_error_detection)
rwlock_error(rwlp, "rwlock_wrlock",
"calling thread owns the writer lock");
error = EDEADLK;
goto out;
}
if (write_lock_try(rwlp, 0))
error = 0;
else if (rwlp->rwlock_type == USYNC_PROCESS)
error = shared_rwlock_lock(rwlp, tsp, WRITE_LOCK);
else
error = rwlock_lock(rwlp, tsp, WRITE_LOCK);
out:
if (error == 0) {
rwlp->rwlock_owner = (uintptr_t)self;
if (rwlp->rwlock_type == USYNC_PROCESS)
rwlp->rwlock_ownerpid = udp->pid;
if (rwsp) {
tdb_incr(rwsp->rw_wrlock);
rwsp->rw_wrlock_begin_hold = gethrtime();
}
DTRACE_PROBE2(plockstat, rw__acquire, rwlp, WRITE_LOCK);
} else {
DTRACE_PROBE3(plockstat, rw__error, rwlp, WRITE_LOCK, error);
}
return (error);
}
#pragma weak pthread_rwlock_wrlock = rw_wrlock
#pragma weak _rw_wrlock = rw_wrlock
int
rw_wrlock(rwlock_t *rwlp)
{
ASSERT(!curthread->ul_critical || curthread->ul_bindflags);
return (rw_wrlock_impl(rwlp, NULL));
}
void
lrw_wrlock(rwlock_t *rwlp)
{
enter_critical(curthread);
(void) rw_wrlock_impl(rwlp, NULL);
}
int
pthread_rwlock_relclockwrlock_np(pthread_rwlock_t *restrict rwlp,
clockid_t clock, const struct timespec *restrict reltime)
{
timespec_t tslocal = *reltime;
int error;
ASSERT(!curthread->ul_critical || curthread->ul_bindflags);
switch (clock) {
case CLOCK_REALTIME:
case CLOCK_HIGHRES:
break;
default:
return (EINVAL);
}
error = rw_wrlock_impl((rwlock_t *)rwlp, &tslocal);
if (error == ETIME)
error = ETIMEDOUT;
return (error);
}
int
pthread_rwlock_reltimedwrlock_np(pthread_rwlock_t *restrict rwlp,
const struct timespec *restrict reltime)
{
return (pthread_rwlock_relclockwrlock_np(rwlp, CLOCK_REALTIME,
reltime));
}
int
pthread_rwlock_clockwrlock(pthread_rwlock_t *rwlp, clockid_t clock,
const timespec_t *abstime)
{
timespec_t tslocal;
int error;
ASSERT(!curthread->ul_critical || curthread->ul_bindflags);
switch (clock) {
case CLOCK_REALTIME:
case CLOCK_HIGHRES:
break;
default:
return (EINVAL);
}
abstime_to_reltime(clock, abstime, &tslocal);
error = rw_wrlock_impl((rwlock_t *)rwlp, &tslocal);
if (error == ETIME)
error = ETIMEDOUT;
return (error);
}
int
pthread_rwlock_timedwrlock(pthread_rwlock_t *rwlp, const timespec_t *abstime)
{
return (pthread_rwlock_clockwrlock(rwlp, CLOCK_REALTIME, abstime));
}
#pragma weak pthread_rwlock_tryrdlock = rw_tryrdlock
int
rw_tryrdlock(rwlock_t *rwlp)
{
ulwp_t *self = curthread;
uberdata_t *udp = self->ul_uberdata;
tdb_rwlock_stats_t *rwsp = RWLOCK_STATS(rwlp, udp);
readlock_t *readlockp;
int error;
ASSERT(!curthread->ul_critical || curthread->ul_bindflags);
if (rwsp)
tdb_incr(rwsp->rw_rdlock_try);
sigoff(self);
readlockp = rwl_entry(rwlp);
if (readlockp->rd_count != 0) {
if (readlockp->rd_count == READ_LOCK_MAX) {
sigon(self);
error = EAGAIN;
goto out;
}
sigon(self);
error = 0;
goto out;
}
sigon(self);
if (read_lock_try(rwlp, 0))
error = 0;
else if (rwlp->rwlock_type == USYNC_PROCESS)
error = shared_rwlock_lock(rwlp, NULL, READ_LOCK_TRY);
else
error = rwlock_lock(rwlp, NULL, READ_LOCK_TRY);
out:
if (error == 0) {
sigoff(self);
rwl_entry(rwlp)->rd_count++;
sigon(self);
DTRACE_PROBE2(plockstat, rw__acquire, rwlp, READ_LOCK);
} else {
if (rwsp)
tdb_incr(rwsp->rw_rdlock_try_fail);
if (error != EBUSY) {
DTRACE_PROBE3(plockstat, rw__error, rwlp, READ_LOCK,
error);
}
}
return (error);
}
#pragma weak pthread_rwlock_trywrlock = rw_trywrlock
int
rw_trywrlock(rwlock_t *rwlp)
{
ulwp_t *self = curthread;
uberdata_t *udp = self->ul_uberdata;
tdb_rwlock_stats_t *rwsp = RWLOCK_STATS(rwlp, udp);
int error;
ASSERT(!self->ul_critical || self->ul_bindflags);
if (rwsp)
tdb_incr(rwsp->rw_wrlock_try);
if (write_lock_try(rwlp, 0))
error = 0;
else if (rwlp->rwlock_type == USYNC_PROCESS)
error = shared_rwlock_lock(rwlp, NULL, WRITE_LOCK_TRY);
else
error = rwlock_lock(rwlp, NULL, WRITE_LOCK_TRY);
if (error == 0) {
rwlp->rwlock_owner = (uintptr_t)self;
if (rwlp->rwlock_type == USYNC_PROCESS)
rwlp->rwlock_ownerpid = udp->pid;
if (rwsp)
rwsp->rw_wrlock_begin_hold = gethrtime();
DTRACE_PROBE2(plockstat, rw__acquire, rwlp, WRITE_LOCK);
} else {
if (rwsp)
tdb_incr(rwsp->rw_wrlock_try_fail);
if (error != EBUSY) {
DTRACE_PROBE3(plockstat, rw__error, rwlp, WRITE_LOCK,
error);
}
}
return (error);
}
#pragma weak pthread_rwlock_unlock = rw_unlock
#pragma weak _rw_unlock = rw_unlock
int
rw_unlock(rwlock_t *rwlp)
{
volatile uint32_t *rwstate = (volatile uint32_t *)&rwlp->rwlock_readers;
uint32_t readers;
ulwp_t *self = curthread;
uberdata_t *udp = self->ul_uberdata;
tdb_rwlock_stats_t *rwsp;
int rd_wr;
readers = *rwstate;
ASSERT_CONSISTENT_STATE(readers);
if (readers & URW_WRITE_LOCKED) {
rd_wr = WRITE_LOCK;
readers = 0;
} else {
rd_wr = READ_LOCK;
readers &= URW_READERS_MASK;
}
if (rd_wr == WRITE_LOCK) {
if (!rw_write_held(rwlp)) {
if (self->ul_error_detection)
rwlock_error(rwlp, "rwlock_unlock",
"writer lock held, "
"but not by the calling thread");
return (EPERM);
}
if ((rwsp = RWLOCK_STATS(rwlp, udp)) != NULL) {
if (rwsp->rw_wrlock_begin_hold)
rwsp->rw_wrlock_hold_time +=
gethrtime() - rwsp->rw_wrlock_begin_hold;
rwsp->rw_wrlock_begin_hold = 0;
}
rwlp->rwlock_owner = 0;
rwlp->rwlock_ownerpid = 0;
} else if (readers > 0) {
readlock_t *readlockp;
sigoff(self);
readlockp = rwl_entry(rwlp);
if (readlockp->rd_count == 0) {
sigon(self);
if (self->ul_error_detection)
rwlock_error(rwlp, "rwlock_unlock",
"readers lock held, "
"but not by the calling thread");
return (EPERM);
}
if (--readlockp->rd_count != 0) {
sigon(self);
goto out;
}
sigon(self);
} else {
if (self->ul_error_detection)
rwlock_error(rwlp, "rwlock_unlock", "lock not owned");
return (EPERM);
}
if (rd_wr == WRITE_LOCK && write_unlock_try(rwlp)) {
;
} else if (rd_wr == READ_LOCK && read_unlock_try(rwlp)) {
;
} else if (rwlp->rwlock_type == USYNC_PROCESS) {
(void) mutex_lock(&rwlp->mutex);
(void) __lwp_rwlock_unlock(rwlp);
(void) mutex_unlock(&rwlp->mutex);
} else {
rw_queue_release(rwlp);
}
out:
DTRACE_PROBE2(plockstat, rw__release, rwlp, rd_wr);
return (0);
}
void
lrw_unlock(rwlock_t *rwlp)
{
(void) rw_unlock(rwlp);
exit_critical(curthread);
}