root/usr/src/uts/intel/asm/cpu.h
/*
 * CDDL HEADER START
 *
 * The contents of this file are subject to the terms of the
 * Common Development and Distribution License (the "License").
 * You may not use this file except in compliance with the License.
 *
 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
 * or http://www.opensolaris.org/os/licensing.
 * See the License for the specific language governing permissions
 * and limitations under the License.
 *
 * When distributing Covered Code, include this CDDL HEADER in each
 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
 * If applicable, add the following below this CDDL HEADER, with the
 * fields enclosed by brackets "[]" replaced with your own identifying
 * information: Portions Copyright [yyyy] [name of copyright owner]
 *
 * CDDL HEADER END
 */
/*
 * Copyright 2009 Sun Microsystems, Inc.  All rights reserved.
 * Use is subject to license terms.
 */
/*
 * Copyright 2019 Joyent, Inc.
 */

#ifndef _ASM_CPU_H
#define _ASM_CPU_H

#include <sys/ccompile.h>

#ifdef  __cplusplus
extern "C" {
#endif

#if !defined(__lint) && defined(__GNUC__)

#if defined(__i386) || defined(__amd64)

extern __GNU_INLINE void
ht_pause(void)
{
        __asm__ __volatile__(
            "pause");
}

/*
 * prefetch 64 bytes
 *
 * prefetch is an SSE extension which is not supported on
 * older 32-bit processors, so define this as a no-op for now
 */

extern __GNU_INLINE void
prefetch_read_many(void *addr)
{
#if defined(__amd64)
        __asm__(
            "prefetcht0 (%0);"
            "prefetcht0 32(%0);"
            : /* no output */
            : "r" (addr));
#endif  /* __amd64 */
}

extern __GNU_INLINE void
prefetch_read_once(void *addr)
{
#if defined(__amd64)
        __asm__(
            "prefetchnta (%0);"
            "prefetchnta 32(%0);"
            : /* no output */
            : "r" (addr));
#endif  /* __amd64 */
}

extern __GNU_INLINE void
prefetch_write_many(void *addr)
{
#if defined(__amd64)
        __asm__(
            "prefetcht0 (%0);"
            "prefetcht0 32(%0);"
            : /* no output */
            : "r" (addr));
#endif  /* __amd64 */
}

extern __GNU_INLINE void
prefetch_write_once(void *addr)
{
#if defined(__amd64)
        __asm__(
            "prefetcht0 (%0);"
            "prefetcht0 32(%0);"
            : /* no output */
            : "r" (addr));
#endif  /* __amd64 */
}

#if !defined(__xpv)

extern __GNU_INLINE void
cli(void)
{
        __asm__ __volatile__(
            "cli" : : : "memory");
}

extern __GNU_INLINE void
sti(void)
{
        __asm__ __volatile__(
            "sti");
}

/*
 * Any newer callers of halt need to make sure that they consider calling
 * x86_md_clear() before calling this to deal with any potential issues with
 * MDS. Because this version of hlt is also used in panic context, we do not
 * unconditionally call x86_md_clear() here and require callers to do so.
 */
extern __GNU_INLINE void
i86_halt(void)
{
        __asm__ __volatile__(
            "sti; hlt");
}

#endif /* !__xpv */

#endif  /* __i386 || defined(__amd64) */

#if defined(__amd64)

extern __GNU_INLINE void
__set_ds(selector_t value)
{
        __asm__ __volatile__(
            "movw       %0, %%ds"
            : /* no output */
            : "r" (value));
}

extern __GNU_INLINE void
__set_es(selector_t value)
{
        __asm__ __volatile__(
            "movw       %0, %%es"
            : /* no output */
            : "r" (value));
}

extern __GNU_INLINE void
__set_fs(selector_t value)
{
        __asm__ __volatile__(
            "movw       %0, %%fs"
            : /* no output */
            : "r" (value));
}

extern __GNU_INLINE void
__set_gs(selector_t value)
{
        __asm__ __volatile__(
            "movw       %0, %%gs"
            : /* no output */
            : "r" (value));
}

#endif  /* __amd64 */

#endif  /* !__lint && __GNUC__ */

#ifdef  __cplusplus
}
#endif

#endif  /* _ASM_CPU_H */