smp_wmb
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* Make pte visible before pmd */
smp_wmb(); /* Make pte visible before pmd */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
#define smp_mb__before_llsc() smp_wmb()
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* goes with smp_mb() in cleanup_single_escalation */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* smp write barrier */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* Makes sure that increment of seq is reflected */
smp_wmb(); /* Makes the value of seq current before we increment */
smp_wmb();
smp_wmb(); /* paired with smp_rmb() in uml_start_secondary() */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* update node_id before generation */
smp_wmb(); /* update node_id before generation */
smp_wmb(); /* update node_id before generation */
smp_wmb();
smp_wmb(); /* node IDs must not be older than generation */
smp_wmb(); /* node IDs must not be older than generation */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* complete the seqlock for execlists_active() */
smp_wmb(); /* notify execlists_active() */
smp_wmb(); /* complete the seqlock */
smp_wmb();
smp_wmb(); /* flush the change in state before reallocation */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* Ensure above write visible before wake */
smp_wmb(); /* Ensure above writes visible before wake */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* see get_txhead() */
smp_wmb();
smp_wmb(); /* make sure errcode is visible first */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* see request builders */
smp_wmb(); /* see qp_set_savail */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
#define journal_entry_set_sector(je, x) do { smp_wmb(); WRITE_ONCE((je)->u.sector, cpu_to_le64(x)); } while (0)
#define journal_entry_set_sector(je, x) do { (je)->u.s.sector_lo = cpu_to_le32(x); smp_wmb(); WRITE_ONCE((je)->u.s.sector_hi, cpu_to_le32((x) >> 32)); } while (0)
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* receive RX msg */
smp_wmb();
smp_wmb(); /* a non-NULL func must hit memory after the key */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* drain writebuffer */
smp_wmb(); /* drain writebuffer */
smp_wmb(); /* drain writebuffer */
smp_wmb(); /* drain writebuffer */
smp_wmb(); /* drain writebuffer */
smp_wmb(); /* drain writebuffer */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* write interrupts_masked before calling napi */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* Ensure tx_tail visible to xmit */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* Pairs with mlx5e_napi_poll. */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* order the skb->data vs. cb->state */
smp_wmb(); /* order the cb->err vs. cb->state */
smp_wmb(); /* order the cb->err vs. cb->state */
smp_wmb();
smp_wmb(); /* Make sure this is seen by ll2_lb_rxq_completion */
smp_wmb(); /* Make sure this is seen by ll2_lb_rxq_completion */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); \
smp_wmb(); \
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* Make these available to spi_finalize_current_message() */
smp_wmb(); /* Matches smp_rmb() in echo_buf(). */
smp_wmb();
smp_wmb(); /* against wdm_read() */
smp_wmb(); /* ordered against wdm_open() */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* All writes before unmask must be visible. */
smp_wmb(); /* Ensure ring contents visible */
smp_wmb();
smp_wmb();
smp_wmb(); /* Set cell state before task state */
smp_wmb(); /* Set rtt before responded. */
smp_wmb(); /* Barrier: set error before flag. */
smp_wmb(); /* Set rtt before responded. */
smp_wmb();
smp_wmb(); /* make event visible before updating tail */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* Write lists before ALL_QUEUED. */
smp_wmb(); /* Write lists before ALL_QUEUED. */
smp_wmb(); /* Write lists before ALL_QUEUED. */
smp_wmb(); /* Write lists before ALL_QUEUED. */
smp_wmb(); /* Write lists before ALL_QUEUED. */
smp_wmb(); /* Write lists before ALL_QUEUED. */
smp_wmb(); /* Write subreq lists before ALL_QUEUED. */
smp_wmb(); /* Write lists before ALL_QUEUED. */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* paired with smp_rmb in xlog_grant_space_left */
#ifndef smp_wmb
#ifndef smp_wmb
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); \
smp_wmb(); \
smp_wmb(); \
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* prior stores before incrementing "sequence" */
smp_wmb(); /* increment "sequence" before following stores */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* B, matches C */
smp_wmb(); /* pairs with the smp_rmb() in handle_swbp() */
smp_wmb();
smp_wmb();
KCSAN_EXPECT_READ_BARRIER(smp_wmb(), false);
KCSAN_EXPECT_WRITE_BARRIER(smp_wmb(), true);
KCSAN_EXPECT_RW_BARRIER(smp_wmb(), true);
KCSAN_CHECK_WRITE_BARRIER(smp_wmb());
KCSAN_CHECK_RW_BARRIER(smp_wmb());
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* Mods to old_rp must follow rcu_assign_pointer() */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); \
smp_wmb();
smp_wmb(); /* func and info before csd. */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* Needed for RCU */
smp_wmb();
smp_wmb(); /* make pte visible before pmd */
smp_wmb(); /* make pte visible before pmd */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* Could be smp_wmb__xxx(before|after)_spin_lock */
smp_wmb(); /* See comment in pmd_install() */
smp_wmb(); /* See comment in pmd_install() */
smp_wmb(); /* See comment in pmd_install() */
smp_wmb(); /* See comment in pmd_install() */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* ensure initialisation is complete before it's called */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* Do we really need it now ? */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* vs rxrpc_process_delayed_final_acks() */
smp_wmb();
smp_wmb(); /* Write data before avail bit */
smp_wmb();
smp_wmb(); /* paired with smp_rmb() in cache_is_valid() */
smp_wmb();
smp_wmb();
smp_wmb(); /* See smp_rmb() in svc_xprt_ready() */
smp_wmb(); /* See smp_rmb() in svc_xprt_ready() */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); /* update buffer data before buffer pointer */
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb(); // Make sure above modifications are seen by i_usx2y_subs_startup()
#ifndef smp_wmb
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();
smp_wmb();