lock1
static time_t lock1(char *, char *);
t = lock1(locktmp, curlock);
nvme_ioctl_lock_t lock1 = *test->cbt_lock1;
if (ioctl(fd1, NVME_IOC_LOCK, &lock1) != 0) {
} else if (lock1.nil_common.nioc_drv_err != test->cbt_ret1) {
lock1.nil_common.nioc_drv_err, test->cbt_ret1);
lockinfo_t *lock1, lockinfo_t *lock2, boolean_t kill_firstborn)
flock_run(styleA, B_TRUE, lock1, &pidA, fdsA);
flock_run(styleB, B_TRUE, lock1, &pidB, fdsB);
assert_write_locked_by(lock1, styleA == LSTYLE_POSIX ? pidA : -1);
assert_write_locked_by(lock1, styleB == LSTYLE_POSIX ? pidB : -1);
lockinfo_t *lock1, lockinfo_t *lock2, boolean_t kill_firstborn)
flock_run(styleB, B_FALSE, lock1, &pidB, fdsB);
flock_run(styleA, B_FALSE, lock1, &pidA, fdsA);
assert_read_locked_by(lock1, styleA == LSTYLE_POSIX ? pidA : -1);
assert_read_locked_by(lock1, styleB == LSTYLE_POSIX ? pidB : -1);
lock_descriptor_t *lock, *lock1;
lock1 = lock->l_next;
lock = lock1;
lock_descriptor_t *lock1, *lock2;
lock1 = flk_get_lock();
COPY(lock1, lock);
lock1->l_start = lock->l_start;
lock1->l_end = request->l_start - 1;
topology[0] = lock1;
lock1 = flk_get_lock();
COPY(lock1, lock);
lock1->l_start = request->l_end + 1;
topology[0] = lock1;
lock1 = flk_get_lock();
COPY(lock1, lock);
lock1->l_start = request->l_end + 1;
topology[0] = lock1;
lock1 = flk_get_lock();
COPY(lock1, lock);
lock1->l_end = request->l_start - 1;
topology[0] = lock1;
lock1 = flk_get_lock();
COPY(lock1, lock);
lock1->l_end = request->l_start - 1;
topology[0] = lock1;
lock_descriptor_t *lock, *lock1;
for (lock1 = lock->l_next; lock1 != ACTIVE_HEAD(gp);
lock1 = lock1->l_next) {
if (lock1->l_vnode == lock->l_vnode) {
if (BLOCKS(lock1, lock)) {
(void *)lock1, (void *)lock);
} else if (BLOCKS(lock, lock1)) {
(void *)lock, (void *)lock1);
lock_descriptor_t *lock1, *lock2;
for (lock1 = SLEEPING_HEAD(gp)->l_next; lock1 != SLEEPING_HEAD(gp);
lock1 = lock1->l_next) {
ASSERT(!IS_BARRIER(lock1));
for (lock2 = lock1->l_next; lock2 != SLEEPING_HEAD(gp);
if (lock1->l_vnode == lock2->l_vnode) {
if (BLOCKS(lock2, lock1)) {
ASSERT(!IS_GRANTED(lock1));
ASSERT(!NOT_BLOCKED(lock1));
path(lock1, lock2);
ASSERT(!IS_BARRIER(lock1));
if (lock1->l_vnode == lock2->l_vnode) {
if (BLOCKS(lock2, lock1)) {
ASSERT(!IS_GRANTED(lock1));
ASSERT(!NOT_BLOCKED(lock1));
path(lock1, lock2);
ep = FIRST_ADJ(lock1);
while (ep != HEAD(lock1)) {
ASSERT(BLOCKS(ep->to_vertex, lock1));
level_two_path(lock_descriptor_t *lock1, lock_descriptor_t *lock2, int no_path)
flk_graph_uncolor(lock1->l_graph);
ep = FIRST_ADJ(lock1);
ASSERT(ep != HEAD(lock1));
while (ep != HEAD(lock1)) {
level_one_path(lock_descriptor_t *lock1, lock_descriptor_t *lock2)
edge_t *ep = FIRST_ADJ(lock1);
while (ep != HEAD(lock1)) {
no_path(lock_descriptor_t *lock1, lock_descriptor_t *lock2)
return (!level_two_path(lock1, lock2, 1));
path(lock_descriptor_t *lock1, lock_descriptor_t *lock2)
if (level_one_path(lock1, lock2)) {
if (level_two_path(lock1, lock2, 0) != 0) {
(void *)lock1, (void *)lock2);
} else if (no_path(lock1, lock2)) {
(void *)lock1, (void *)lock2);
#define SAME_OWNER(lock1, lock2) \
(((lock1)->l_flock.l_pid == (lock2)->l_flock.l_pid) && \
((lock1)->l_flock.l_sysid == (lock2)->l_flock.l_sysid) && \
((lock1)->l_ofd == (lock2)->l_ofd))
#define OVERLAP(lock1, lock2) \
(((lock1)->l_start <= (lock2)->l_start && \
(lock2)->l_start <= (lock1)->l_end) || \
((lock2)->l_start <= (lock1)->l_start && \
(lock1)->l_start <= (lock2)->l_end))
#define BLOCKS(lock1, lock2) (!SAME_OWNER((lock1), (lock2)) && \
(((lock1)->l_type == F_WRLCK) || \
OVERLAP((lock1), (lock2)))
#define COVERS(lock1, lock2) \
(((lock1)->l_start <= (lock2)->l_start) && \
((lock1)->l_end >= (lock2)->l_end))
#define COPY(lock1, lock2) \
(lock1)->l_graph = (lock2)->l_graph; \
(lock1)->l_vnode = (lock2)->l_vnode; \
(lock1)->l_type = (lock2)->l_type; \
(lock1)->l_state = (lock2)->l_state; \
(lock1)->l_start = (lock2)->l_start; \
(lock1)->l_end = (lock2)->l_end; \
(lock1)->l_flock = (lock2)->l_flock; \
(lock1)->l_zoneid = (lock2)->l_zoneid; \
(lock1)->pvertex = (lock2)->pvertex; \