smp_rmb
smp_rmb(); /* see above */
smp_rmb(); /* see above */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* See __yield_to_locked_owner comment */
smp_rmb(); /* acquire barrier for the mcs lock */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* Finish all reads before we return seq */
smp_rmb(); /* Finish all reads before checking the value of seq */
smp_rmb();
smp_rmb(); /* paired with smp_wmb() in __cpu_up() */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* node_id vs. generation */
smp_rmb(); /* sbp2_map_scatterlist looks at tgt->address_high */
smp_rmb(); /* node IDs must not be older than generation */
smp_rmb(); /* get current card generation */
smp_rmb(); /* generation may have changed */
smp_rmb(); /* node_id must not be older than generation */
smp_rmb(); /* node IDs must not be older than generation */
smp_rmb(); /* get current card generation */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* pairs with overwrite protection in process_csb() */
smp_rmb(); /* and complete the seqlock retry */
smp_rmb(); /* serialise with add_active_barriers */
smp_rmb(); /* serialise with add_active_barriers */
smp_rmb(); /* serialise with add_active_barriers */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* prevent speculative reads of dma'ed hdrq */
smp_rmb();
smp_rmb(); /* prevent speculative reads of dma'ed hdrq */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* Make sure we access most actual mddev properties */
smp_rmb(); /* Make sure we access most recent reshape position */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* node_id vs. generation */
smp_rmb(); /* node_id vs. generation */
smp_rmb(); /* RX msg has been received */
smp_rmb(); /* ensure later readers know we're not busy */
smp_rmb(); /* ensure later readers know we're busy */
smp_rmb(); /* ensure later readers know we're not busy */
smp_rmb();
smp_rmb(); /* ensure a fresh copy of part_uv->flags */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* make sure interrupts_masked is read */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* read error code after completion */
smp_rmb(); /* verify reading after completion */
smp_rmb(); /* read error code after completion */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* prevent any other reads prior to eop_desc */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* Ensure the latest updates are visible */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* Pairs with mlx5e_qos_alloc_queues. */
smp_rmb();
smp_rmb();
smp_rmb(); /* pairs with smp_wmb() after data is written */
smp_rmb(); /* pairs with wmb after data is written */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* Paired with netif_tx_stop_queue set_bit */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* Matches smp_wmb() in add_echo_byte(). */
smp_rmb();
smp_rmb(); /* against wdm_in_callback() */
smp_rmb(); /* ordered against wdm_wwan_port_stop() */
smp_rmb();
smp_rmb(); /* paired with atomic_inc() and atomic_dec_return() */
smp_rmb();
smp_rmb(); /* Ensure that we see the port before we copy it. */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* Read counters after IN_PROGRESS flag. */
smp_rmb(); /* Read ALL_QUEUED before subreq lists. */
smp_rmb();
smp_rmb(); /* Read counters after I-P flag. */
smp_rmb(); /* Read ALL_QUEUED before lists. */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* pairs with smp_wmb() below */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* paired with smp_wmb in xlog_cil_ail_insert() */
smp_rmb();
#ifndef smp_rmb
#ifndef smp_rmb
#define smp_acquire__after_ctrl_dep() smp_rmb()
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); \
smp_rmb(); /* coupled with smp_wmb() in vlan_dev_set_egress_priority() */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
KCSAN_EXPECT_READ_BARRIER(smp_rmb(), true);
KCSAN_EXPECT_WRITE_BARRIER(smp_rmb(), false);
KCSAN_EXPECT_RW_BARRIER(smp_rmb(), true);
KCSAN_CHECK_READ_BARRIER(smp_rmb());
KCSAN_CHECK_RW_BARRIER(smp_rmb());
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* LMM(prb_first_seq:C) */
smp_rmb(); /* LMM(desc_read:B) */
smp_rmb(); /* LMM(desc_read:D) */
smp_rmb(); /* LMM(data_push_tail:B) */
smp_rmb(); /* LMM(desc_push_tail:C) */
smp_rmb(); /* LMM(desc_reserve:B) */
smp_rmb();
smp_rmb();
smp_rmb(); // Order CT state and *vp reads.
smp_rmb(); // Order *vp read and CT state re-read.
smp_rmb(); /* Pick up ->gp_seq first... */
smp_rmb(); /* ...then ->jiffies_stall before the rest... */
smp_rmb(); /* ...and finally ->gp_start before ->gp_seq again. */
smp_rmb(); /* see try_to_wake_up() */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); \
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* Must read tx_psock before tx_wait */
smp_rmb(); /* Read tx_psock before tx_wait */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* Read avail bits before accessing data. */
smp_rmb(); /* coupled with smp_wmb() in tcf_gact_init() */
smp_rmb(); /* coupled with smp_wmb() in tcf_gact_init() */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); \
smp_rmb(); /* paired with atomic_inc() and atomic_dec_return() */
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb(); /* node_id vs. generation */
smp_rmb(); /* node_id vs. generation */
smp_rmb();
smp_rmb(); /* node id vs. generation */
smp_rmb(); /* node_id vs. generation */
smp_rmb(); /* node_id vs. generation */
smp_rmb(); /* node_id vs. generation */
smp_rmb();
#ifndef smp_rmb
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
#define smp_acquire__after_ctrl_dep() ({ smp_rmb(); })
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();
smp_rmb();