hash_entry
nscd_hash_t *idx_p, *hash_entry;
hash_entry = (nscd_hash_t *)entry;
hash_entry->prev_p = prev_p;
db->hash_tbl_p[i] = hash_entry;
prev_p->next_p = hash_entry;
hash_entry->next_p = next_p;
next_p->prev_p = hash_entry;
hash_entry(_cache, &cache[sshdr->sh_link], hsecname,
hash_entry(_cache, &cache[sshdr->sh_link],
hash_entry *tmp;
hash_entry * tmp, * prev = NULL;
hash_entry *tmp = tbl->start;
hash_entry *tmp = tbl->start;
hash_entry * tmp = tbl->start, * prev;
ptr->table = malloc((size_t)(sizeof (hash_entry *) * size));
(void) memset((char *)ptr->table, 0, sizeof (hash_entry *) * size);
ptr->table = malloc(sizeof (hash_entry *) * size);
(void) memset((char *)ptr->table, 0, sizeof (hash_entry *) * size);
hash_entry *tmp, *new;
struct hash_entry *next_entry;
struct hash_entry *right_entry;
struct hash_entry *left_entry;
hash_entry **table;
hash_entry *start;
(void) pthread_mutex_unlock(&hash_entry->sip_hash_mutex);
sip_hash_t *hash_entry;
hash_entry = &sip_hash[count];
(void) pthread_mutex_lock(&hash_entry->sip_hash_mutex);
tmp = hash_entry->hash_head;
for (hcount = 0; hcount < hash_entry->hash_count; hcount++) {
(void) pthread_mutex_unlock(&hash_entry->sip_hash_mutex);
sip_hash_t *hash_entry;
hash_entry = &sip_hash[hindex];
(void) pthread_mutex_lock(&hash_entry->sip_hash_mutex);
tmp = hash_entry->hash_head;
for (count = 0; count < hash_entry->hash_count; count++) {
if (tmp == hash_entry->hash_head) {
hash_entry->hash_head = tmp->next_obj;
assert(hash_entry->hash_tail ==
hash_entry->hash_head);
hash_entry->hash_head = NULL;
hash_entry->hash_tail = NULL;
assert(hash_entry->hash_tail == tmp);
hash_entry->hash_tail =
hash_entry->hash_count--;
&hash_entry->sip_hash_mutex);
&hash_entry->sip_hash_mutex);
(void) pthread_mutex_unlock(&hash_entry->sip_hash_mutex);
sip_hash_t *hash_entry;
hash_entry = &sip_hash[hindex];
(void) pthread_mutex_lock(&hash_entry->sip_hash_mutex);
if (hash_entry->hash_count == 0) {
assert(hash_entry->hash_head == NULL);
assert(hash_entry->hash_tail == NULL);
hash_entry->hash_head = new_obj;
hash_entry->hash_tail->next_obj = new_obj;
new_obj->prev_obj = hash_entry->hash_tail;
hash_entry->hash_tail = new_obj;
hash_entry->hash_count++;
(void) pthread_mutex_unlock(&hash_entry->sip_hash_mutex);
sip_hash_t *hash_entry;
hash_entry = &sip_hash[hindex];
(void) pthread_mutex_lock(&hash_entry->sip_hash_mutex);
tmp = hash_entry->hash_head;
for (count = 0; count < hash_entry->hash_count; count++) {
&hash_entry->sip_hash_mutex);
lm_searcher_hash_entry_t *hash_entry = NULL;
hash_entry = &context->searcher_hash.searcher_table[h_val];
temp_ipv6 = hash_entry->num_ipv6 + (is_ipv4 ? 0 : 1);
temp_ipv4 = hash_entry->num_ipv4 + is_ipv4;
if (temp_depth_ipv4 < hash_entry->depth_ipv4) {
temp_depth_ipv4 = hash_entry->depth_ipv4;
hash_entry->num_ipv6 = temp_ipv6;
hash_entry->num_ipv4 = temp_ipv4;
hash_entry->depth_ipv4 = temp_depth_ipv4;
if (context->searcher_hash.hash_depth_reached < hash_entry->depth_ipv4 + hash_entry->num_ipv6) {
context->searcher_hash.hash_depth_reached = hash_entry->depth_ipv4 + hash_entry->num_ipv6;
lm_searcher_hash_entry_t *hash_entry = NULL;
hash_entry = &context->searcher_hash.searcher_table[h_val];
DbgBreakIf(!hash_entry->num_ipv6);
hash_entry->num_ipv6--;
DbgBreakIf(!hash_entry->num_ipv4);
hash_entry->num_ipv4--;
if (hash_entry->num_ipv4 < hash_entry->depth_ipv4)
hash_entry->depth_ipv4 = hash_entry->num_ipv4;