walk
walk:
goto walk;
walk(BIO_next, start, &end, &len);
walk(BIO_prev, end, &start, &len);
rbnode_type* walk, *next;
walk = rbtree_next(&el->node);
while(walk && walk != RBTREE_NULL) {
cur = (struct val_neg_data*)walk;
walk = rbtree_next(walk);
next = rbtree_next(walk);
walk = next;
struct canon_rr* walk;
RBTREE_FOR(walk, struct canon_rr*, (*sortree)) {
+ d->rr_len[walk->rr_idx]) {
sldns_buffer_write(buf, d->rr_data[walk->rr_idx],
d->rr_len[walk->rr_idx]);
canonicalize_rdata(buf, k, d->rr_len[walk->rr_idx]);
struct canon_rr* walk;
RBTREE_FOR(walk, struct canon_rr*, sortree) {
+ d->rr_len[walk->rr_idx]) {
sldns_buffer_write_u32(buf, d->rr_ttl[walk->rr_idx]);
sldns_buffer_write(buf, d->rr_data[walk->rr_idx],
d->rr_len[walk->rr_idx]);
canonicalize_rdata(buf, k, d->rr_len[walk->rr_idx]);
s64 (*process_bo)(struct ttm_lru_walk *walk, struct ttm_buffer_object *bo);
s64 ttm_lru_walk_for_evict(struct ttm_lru_walk *walk, struct ttm_device *bdev,
struct ttm_lru_walk walk;
ttm_bo_swapout_cb(struct ttm_lru_walk *walk, struct ttm_buffer_object *bo)
container_of(walk, typeof(*swapout_walk), walk);
struct ttm_operation_ctx *ctx = walk->arg.ctx;
.walk = {
return ttm_lru_walk_for_evict(&swapout_walk.walk, bdev, man, target);
struct ttm_lru_walk walk;
static s64 ttm_bo_evict_cb(struct ttm_lru_walk *walk, struct ttm_buffer_object *bo)
container_of(walk, typeof(*evict_walk), walk);
lret = ttm_bo_wait_ctx(bo, walk->arg.ctx);
lret = ttm_bo_evict(bo, walk->arg.ctx);
.walk = {
evict_walk.walk.arg.trylock_only = true;
lret = ttm_lru_walk_for_evict(&evict_walk.walk, bdev, man, 1);
lret = ttm_lru_walk_for_evict(&evict_walk.walk, bdev, man, 1);
evict_walk.walk.arg.trylock_only = false;
evict_walk.walk.arg.ticket = ticket;
lret = ttm_lru_walk_for_evict(&evict_walk.walk, bdev, man, 1);
s64 ttm_lru_walk_for_evict(struct ttm_lru_walk *walk, struct ttm_device *bdev,
ttm_bo_lru_for_each_reserved_guarded(&cursor, man, &walk->arg, bo) {
lret = walk->ops->process_bo(walk, bo);
uvm_coredump_walk_cb *walk, void *cookie)
error = (*walk)(start, realend, end, prot,
error = (*walk)(start, realend, entry_end, prot, 0, nsegment, cookie);
uvm_coredump_walk_cb *walk, void *cookie)
walk, cookie);
error = (*walk)(entry->start, end, end, entry->protection,
domain_type* walk;
walk = zone->apex;
while(walk && domain_is_subdomain(walk, zone->apex)) {
if(walk->nsec3) {
if(nsec3_condition_hash(walk, zone)) {
walk->nsec3->nsec3_node.key = NULL;
walk->nsec3->nsec3_cover = NULL;
walk->nsec3->nsec3_wcard_child_cover = NULL;
walk->nsec3->nsec3_is_exact = 0;
if (walk->nsec3->hash_wc) {
walk->nsec3->hash_wc,
walk->nsec3->hash_wc = NULL;
if(nsec3_condition_dshash(walk, zone)) {
walk->nsec3->nsec3_ds_parent_cover = NULL;
walk->nsec3->nsec3_ds_parent_is_exact = 0;
if (walk->nsec3->ds_parent_hash) {
walk->nsec3->ds_parent_hash,
walk->nsec3->ds_parent_hash = NULL;
walk = domain_next(walk);
domain_type* walk;
for(walk=zone->apex; walk && domain_is_subdomain(walk, zone->apex);
walk = domain_next(walk)) {
if(nsec3_in_chain_count(walk, zone) != 0) {
nsec3_precompile_nsec3rr(db, walk, zone);
for(walk=zone->apex; walk && domain_is_subdomain(walk, zone->apex);
walk = domain_next(walk)) {
if(nsec3_condition_hash(walk, zone)) {
nsec3_precompile_domain(db, walk, zone, tmpregion);
if(nsec3_condition_dshash(walk, zone))
nsec3_precompile_domain_ds(db, walk, zone);
domain_type* walk;
walk = query->zone->apex;
while(walk) {
if(walk->nsec3 && walk->nsec3->nsec3_cover == cover) {
"%s %s", domain_to_string(walk),
walk->nsec3->nsec3_is_exact?"exact":"no_exact_hash_match");
if(walk->nsec3->nsec3_is_exact)
if(walk->nsec3 && walk->nsec3->nsec3_ds_parent_cover == cover) {
"%s %s", domain_to_string(walk),
walk->nsec3->nsec3_ds_parent_is_exact?"exact":"no_exact_hash_match");
if(walk->nsec3->nsec3_ds_parent_is_exact)
walk = domain_next(walk);
(void)sldns_wire2str_rr_scan(&walk, &walk_len, &snull, &sl,
uint8_t* walk = p;
walk += LDNS_HEADER_SIZE;
(void)sldns_wire2str_rrquestion_scan(&walk, &walk_len,
(void)sldns_wire2str_rr_scan(&walk, &walk_len, &snull, &sl,
uint8_t* dstart = walk;
(void)sldns_wire2str_rr_scan(&walk, &walk_len, &snull, &sl,
uint8_t* walk = pkt;
if(!pkt_find_edns_opt(&walk, &walk_len)) {
edns_bits = sldns_read_uint16(walk+4);
uint8_t* walk = pkt;
walk += LDNS_HEADER_SIZE;
(void)sldns_wire2str_rrquestion_scan(&walk, &walk_len,
uint8_t* dstart = walk;
rbnode_type* walk;
walk = rbtree_first(&neg->tree); /* first highest parent, big count */
z = (struct val_neg_zone*)walk;
walk = rbtree_first(&neg->tree);
z = (struct val_neg_zone*)walk;
while(i!=n+1 && walk && walk != RBTREE_NULL && !z->in_use) {
walk = rbtree_next(walk);
z = (struct val_neg_zone*)walk;
if(!walk || walk == RBTREE_NULL) {
walk = rbtree_first(&z->tree); /* first is highest parent */
d = (struct val_neg_data*)walk;
walk = rbtree_first(&z->tree);
d = (struct val_neg_data*)walk;
while(i!=n+1 && walk && walk != RBTREE_NULL && !d->in_use) {
walk = rbtree_next(walk);
d = (struct val_neg_data*)walk;
if(!walk || walk == RBTREE_NULL) {
rbnode_type* walk, *next;
walk = rbtree_next(&el->node);
while(walk && walk != RBTREE_NULL) {
cur = (struct val_neg_data*)walk;
walk = rbtree_next(walk);
next = rbtree_next(walk);
walk = next;
struct canon_rr* walk;
RBTREE_FOR(walk, struct canon_rr*, (*sortree)) {
+ d->rr_len[walk->rr_idx]) {
sldns_buffer_write(buf, d->rr_data[walk->rr_idx],
d->rr_len[walk->rr_idx]);
canonicalize_rdata(buf, k, d->rr_len[walk->rr_idx]);
struct canon_rr* walk;
RBTREE_FOR(walk, struct canon_rr*, sortree) {
+ d->rr_len[walk->rr_idx]) {
sldns_buffer_write_u32(buf, d->rr_ttl[walk->rr_idx]);
sldns_buffer_write(buf, d->rr_data[walk->rr_idx],
d->rr_len[walk->rr_idx]);
canonicalize_rdata(buf, k, d->rr_len[walk->rr_idx]);