root/arch/powerpc/lib/crtsavres.S
/*
 * Special support for eabi and SVR4
 *
 *   Copyright (C) 1995, 1996, 1998, 2000, 2001 Free Software Foundation, Inc.
 *   Copyright 2008 Freescale Semiconductor, Inc.
 *   Written By Michael Meissner
 *
 * Based on gcc/config/rs6000/crtsavres.asm from gcc
 * 64 bit additions from reading the PPC elf64abi document.
 *
 * This file is free software; you can redistribute it and/or modify it
 * under the terms of the GNU General Public License as published by the
 * Free Software Foundation; either version 2, or (at your option) any
 * later version.
 *
 * In addition to the permissions in the GNU General Public License, the
 * Free Software Foundation gives you unlimited permission to link the
 * compiled version of this file with other programs, and to distribute
 * those programs without any restriction coming from the use of this
 * file.  (The General Public License restrictions do apply in other
 * respects; for example, they cover modification of the file, and
 * distribution when not linked into another program.)
 *
 * This file is distributed in the hope that it will be useful, but
 * WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 * General Public License for more details.
 *
 * You should have received a copy of the GNU General Public License
 * along with this program; see the file COPYING.  If not, write to
 * the Free Software Foundation, 51 Franklin Street, Fifth Floor,
 * Boston, MA 02110-1301, USA.
 *
 *    As a special exception, if you link this library with files
 *    compiled with GCC to produce an executable, this does not cause
 *    the resulting executable to be covered by the GNU General Public License.
 *    This exception does not however invalidate any other reasons why
 *    the executable file might be covered by the GNU General Public License.
 */

#include <asm/ppc_asm.h>

        .file   "crtsavres.S"

#ifdef CONFIG_CC_OPTIMIZE_FOR_SIZE

        .section ".text"

#ifndef __powerpc64__

/* Routines for saving integer registers, called by the compiler.  */
/* Called with r11 pointing to the stack header word of the caller of the */
/* function, just beyond the end of the integer save area.  */

_GLOBAL(_savegpr_14)
_GLOBAL(_save32gpr_14)
        stw     14,-72(11)      /* save gp registers */
_GLOBAL(_savegpr_15)
_GLOBAL(_save32gpr_15)
        stw     15,-68(11)
_GLOBAL(_savegpr_16)
_GLOBAL(_save32gpr_16)
        stw     16,-64(11)
_GLOBAL(_savegpr_17)
_GLOBAL(_save32gpr_17)
        stw     17,-60(11)
_GLOBAL(_savegpr_18)
_GLOBAL(_save32gpr_18)
        stw     18,-56(11)
_GLOBAL(_savegpr_19)
_GLOBAL(_save32gpr_19)
        stw     19,-52(11)
_GLOBAL(_savegpr_20)
_GLOBAL(_save32gpr_20)
        stw     20,-48(11)
_GLOBAL(_savegpr_21)
_GLOBAL(_save32gpr_21)
        stw     21,-44(11)
_GLOBAL(_savegpr_22)
_GLOBAL(_save32gpr_22)
        stw     22,-40(11)
_GLOBAL(_savegpr_23)
_GLOBAL(_save32gpr_23)
        stw     23,-36(11)
_GLOBAL(_savegpr_24)
_GLOBAL(_save32gpr_24)
        stw     24,-32(11)
_GLOBAL(_savegpr_25)
_GLOBAL(_save32gpr_25)
        stw     25,-28(11)
_GLOBAL(_savegpr_26)
_GLOBAL(_save32gpr_26)
        stw     26,-24(11)
_GLOBAL(_savegpr_27)
_GLOBAL(_save32gpr_27)
        stw     27,-20(11)
_GLOBAL(_savegpr_28)
_GLOBAL(_save32gpr_28)
        stw     28,-16(11)
_GLOBAL(_savegpr_29)
_GLOBAL(_save32gpr_29)
        stw     29,-12(11)
_GLOBAL(_savegpr_30)
_GLOBAL(_save32gpr_30)
        stw     30,-8(11)
_GLOBAL(_savegpr_31)
_GLOBAL(_save32gpr_31)
        stw     31,-4(11)
        blr

/* Routines for restoring integer registers, called by the compiler.  */
/* Called with r11 pointing to the stack header word of the caller of the */
/* function, just beyond the end of the integer restore area.  */

_GLOBAL(_restgpr_14)
_GLOBAL(_rest32gpr_14)
        lwz     14,-72(11)      /* restore gp registers */
_GLOBAL(_restgpr_15)
_GLOBAL(_rest32gpr_15)
        lwz     15,-68(11)
_GLOBAL(_restgpr_16)
_GLOBAL(_rest32gpr_16)
        lwz     16,-64(11)
_GLOBAL(_restgpr_17)
_GLOBAL(_rest32gpr_17)
        lwz     17,-60(11)
_GLOBAL(_restgpr_18)
_GLOBAL(_rest32gpr_18)
        lwz     18,-56(11)
_GLOBAL(_restgpr_19)
_GLOBAL(_rest32gpr_19)
        lwz     19,-52(11)
_GLOBAL(_restgpr_20)
_GLOBAL(_rest32gpr_20)
        lwz     20,-48(11)
_GLOBAL(_restgpr_21)
_GLOBAL(_rest32gpr_21)
        lwz     21,-44(11)
_GLOBAL(_restgpr_22)
_GLOBAL(_rest32gpr_22)
        lwz     22,-40(11)
_GLOBAL(_restgpr_23)
_GLOBAL(_rest32gpr_23)
        lwz     23,-36(11)
_GLOBAL(_restgpr_24)
_GLOBAL(_rest32gpr_24)
        lwz     24,-32(11)
_GLOBAL(_restgpr_25)
_GLOBAL(_rest32gpr_25)
        lwz     25,-28(11)
_GLOBAL(_restgpr_26)
_GLOBAL(_rest32gpr_26)
        lwz     26,-24(11)
_GLOBAL(_restgpr_27)
_GLOBAL(_rest32gpr_27)
        lwz     27,-20(11)
_GLOBAL(_restgpr_28)
_GLOBAL(_rest32gpr_28)
        lwz     28,-16(11)
_GLOBAL(_restgpr_29)
_GLOBAL(_rest32gpr_29)
        lwz     29,-12(11)
_GLOBAL(_restgpr_30)
_GLOBAL(_rest32gpr_30)
        lwz     30,-8(11)
_GLOBAL(_restgpr_31)
_GLOBAL(_rest32gpr_31)
        lwz     31,-4(11)
        blr

/* Routines for restoring integer registers, called by the compiler.  */
/* Called with r11 pointing to the stack header word of the caller of the */
/* function, just beyond the end of the integer restore area.  */

_GLOBAL(_restgpr_14_x)
_GLOBAL(_rest32gpr_14_x)
        lwz     14,-72(11)      /* restore gp registers */
_GLOBAL(_restgpr_15_x)
_GLOBAL(_rest32gpr_15_x)
        lwz     15,-68(11)
_GLOBAL(_restgpr_16_x)
_GLOBAL(_rest32gpr_16_x)
        lwz     16,-64(11)
_GLOBAL(_restgpr_17_x)
_GLOBAL(_rest32gpr_17_x)
        lwz     17,-60(11)
_GLOBAL(_restgpr_18_x)
_GLOBAL(_rest32gpr_18_x)
        lwz     18,-56(11)
_GLOBAL(_restgpr_19_x)
_GLOBAL(_rest32gpr_19_x)
        lwz     19,-52(11)
_GLOBAL(_restgpr_20_x)
_GLOBAL(_rest32gpr_20_x)
        lwz     20,-48(11)
_GLOBAL(_restgpr_21_x)
_GLOBAL(_rest32gpr_21_x)
        lwz     21,-44(11)
_GLOBAL(_restgpr_22_x)
_GLOBAL(_rest32gpr_22_x)
        lwz     22,-40(11)
_GLOBAL(_restgpr_23_x)
_GLOBAL(_rest32gpr_23_x)
        lwz     23,-36(11)
_GLOBAL(_restgpr_24_x)
_GLOBAL(_rest32gpr_24_x)
        lwz     24,-32(11)
_GLOBAL(_restgpr_25_x)
_GLOBAL(_rest32gpr_25_x)
        lwz     25,-28(11)
_GLOBAL(_restgpr_26_x)
_GLOBAL(_rest32gpr_26_x)
        lwz     26,-24(11)
_GLOBAL(_restgpr_27_x)
_GLOBAL(_rest32gpr_27_x)
        lwz     27,-20(11)
_GLOBAL(_restgpr_28_x)
_GLOBAL(_rest32gpr_28_x)
        lwz     28,-16(11)
_GLOBAL(_restgpr_29_x)
_GLOBAL(_rest32gpr_29_x)
        lwz     29,-12(11)
_GLOBAL(_restgpr_30_x)
_GLOBAL(_rest32gpr_30_x)
        lwz     30,-8(11)
_GLOBAL(_restgpr_31_x)
_GLOBAL(_rest32gpr_31_x)
        lwz     0,4(11)
        lwz     31,-4(11)
        mtlr    0
        mr      1,11
        blr

#ifdef CONFIG_ALTIVEC
/* Called with r0 pointing just beyond the end of the vector save area.  */

_GLOBAL(_savevr_20)
        li      r11,-192
        stvx    v20,r11,r0
_GLOBAL(_savevr_21)
        li      r11,-176
        stvx    v21,r11,r0
_GLOBAL(_savevr_22)
        li      r11,-160
        stvx    v22,r11,r0
_GLOBAL(_savevr_23)
        li      r11,-144
        stvx    v23,r11,r0
_GLOBAL(_savevr_24)
        li      r11,-128
        stvx    v24,r11,r0
_GLOBAL(_savevr_25)
        li      r11,-112
        stvx    v25,r11,r0
_GLOBAL(_savevr_26)
        li      r11,-96
        stvx    v26,r11,r0
_GLOBAL(_savevr_27)
        li      r11,-80
        stvx    v27,r11,r0
_GLOBAL(_savevr_28)
        li      r11,-64
        stvx    v28,r11,r0
_GLOBAL(_savevr_29)
        li      r11,-48
        stvx    v29,r11,r0
_GLOBAL(_savevr_30)
        li      r11,-32
        stvx    v30,r11,r0
_GLOBAL(_savevr_31)
        li      r11,-16
        stvx    v31,r11,r0
        blr

_GLOBAL(_restvr_20)
        li      r11,-192
        lvx     v20,r11,r0
_GLOBAL(_restvr_21)
        li      r11,-176
        lvx     v21,r11,r0
_GLOBAL(_restvr_22)
        li      r11,-160
        lvx     v22,r11,r0
_GLOBAL(_restvr_23)
        li      r11,-144
        lvx     v23,r11,r0
_GLOBAL(_restvr_24)
        li      r11,-128
        lvx     v24,r11,r0
_GLOBAL(_restvr_25)
        li      r11,-112
        lvx     v25,r11,r0
_GLOBAL(_restvr_26)
        li      r11,-96
        lvx     v26,r11,r0
_GLOBAL(_restvr_27)
        li      r11,-80
        lvx     v27,r11,r0
_GLOBAL(_restvr_28)
        li      r11,-64
        lvx     v28,r11,r0
_GLOBAL(_restvr_29)
        li      r11,-48
        lvx     v29,r11,r0
_GLOBAL(_restvr_30)
        li      r11,-32
        lvx     v30,r11,r0
_GLOBAL(_restvr_31)
        li      r11,-16
        lvx     v31,r11,r0
        blr

#endif /* CONFIG_ALTIVEC */

#else /* CONFIG_PPC64 */

.globl  _savegpr0_14
_savegpr0_14:
        std     r14,-144(r1)
.globl  _savegpr0_15
_savegpr0_15:
        std     r15,-136(r1)
.globl  _savegpr0_16
_savegpr0_16:
        std     r16,-128(r1)
.globl  _savegpr0_17
_savegpr0_17:
        std     r17,-120(r1)
.globl  _savegpr0_18
_savegpr0_18:
        std     r18,-112(r1)
.globl  _savegpr0_19
_savegpr0_19:
        std     r19,-104(r1)
.globl  _savegpr0_20
_savegpr0_20:
        std     r20,-96(r1)
.globl  _savegpr0_21
_savegpr0_21:
        std     r21,-88(r1)
.globl  _savegpr0_22
_savegpr0_22:
        std     r22,-80(r1)
.globl  _savegpr0_23
_savegpr0_23:
        std     r23,-72(r1)
.globl  _savegpr0_24
_savegpr0_24:
        std     r24,-64(r1)
.globl  _savegpr0_25
_savegpr0_25:
        std     r25,-56(r1)
.globl  _savegpr0_26
_savegpr0_26:
        std     r26,-48(r1)
.globl  _savegpr0_27
_savegpr0_27:
        std     r27,-40(r1)
.globl  _savegpr0_28
_savegpr0_28:
        std     r28,-32(r1)
.globl  _savegpr0_29
_savegpr0_29:
        std     r29,-24(r1)
.globl  _savegpr0_30
_savegpr0_30:
        std     r30,-16(r1)
.globl  _savegpr0_31
_savegpr0_31:
        std     r31,-8(r1)
        std     r0,16(r1)
        blr

.globl  _restgpr0_14
_restgpr0_14:
        ld      r14,-144(r1)
.globl  _restgpr0_15
_restgpr0_15:
        ld      r15,-136(r1)
.globl  _restgpr0_16
_restgpr0_16:
        ld      r16,-128(r1)
.globl  _restgpr0_17
_restgpr0_17:
        ld      r17,-120(r1)
.globl  _restgpr0_18
_restgpr0_18:
        ld      r18,-112(r1)
.globl  _restgpr0_19
_restgpr0_19:
        ld      r19,-104(r1)
.globl  _restgpr0_20
_restgpr0_20:
        ld      r20,-96(r1)
.globl  _restgpr0_21
_restgpr0_21:
        ld      r21,-88(r1)
.globl  _restgpr0_22
_restgpr0_22:
        ld      r22,-80(r1)
.globl  _restgpr0_23
_restgpr0_23:
        ld      r23,-72(r1)
.globl  _restgpr0_24
_restgpr0_24:
        ld      r24,-64(r1)
.globl  _restgpr0_25
_restgpr0_25:
        ld      r25,-56(r1)
.globl  _restgpr0_26
_restgpr0_26:
        ld      r26,-48(r1)
.globl  _restgpr0_27
_restgpr0_27:
        ld      r27,-40(r1)
.globl  _restgpr0_28
_restgpr0_28:
        ld      r28,-32(r1)
.globl  _restgpr0_29
_restgpr0_29:
        ld      r0,16(r1)
        ld      r29,-24(r1)
        mtlr    r0
        ld      r30,-16(r1)
        ld      r31,-8(r1)
        blr

.globl  _restgpr0_30
_restgpr0_30:
        ld      r30,-16(r1)
.globl  _restgpr0_31
_restgpr0_31:
        ld      r0,16(r1)
        ld      r31,-8(r1)
        mtlr    r0
        blr

#ifdef CONFIG_ALTIVEC
/* Called with r0 pointing just beyond the end of the vector save area.  */

.globl  _savevr_20
_savevr_20:
        li      r12,-192
        stvx    v20,r12,r0
.globl  _savevr_21
_savevr_21:
        li      r12,-176
        stvx    v21,r12,r0
.globl  _savevr_22
_savevr_22:
        li      r12,-160
        stvx    v22,r12,r0
.globl  _savevr_23
_savevr_23:
        li      r12,-144
        stvx    v23,r12,r0
.globl  _savevr_24
_savevr_24:
        li      r12,-128
        stvx    v24,r12,r0
.globl  _savevr_25
_savevr_25:
        li      r12,-112
        stvx    v25,r12,r0
.globl  _savevr_26
_savevr_26:
        li      r12,-96
        stvx    v26,r12,r0
.globl  _savevr_27
_savevr_27:
        li      r12,-80
        stvx    v27,r12,r0
.globl  _savevr_28
_savevr_28:
        li      r12,-64
        stvx    v28,r12,r0
.globl  _savevr_29
_savevr_29:
        li      r12,-48
        stvx    v29,r12,r0
.globl  _savevr_30
_savevr_30:
        li      r12,-32
        stvx    v30,r12,r0
.globl  _savevr_31
_savevr_31:
        li      r12,-16
        stvx    v31,r12,r0
        blr

.globl  _restvr_20
_restvr_20:
        li      r12,-192
        lvx     v20,r12,r0
.globl  _restvr_21
_restvr_21:
        li      r12,-176
        lvx     v21,r12,r0
.globl  _restvr_22
_restvr_22:
        li      r12,-160
        lvx     v22,r12,r0
.globl  _restvr_23
_restvr_23:
        li      r12,-144
        lvx     v23,r12,r0
.globl  _restvr_24
_restvr_24:
        li      r12,-128
        lvx     v24,r12,r0
.globl  _restvr_25
_restvr_25:
        li      r12,-112
        lvx     v25,r12,r0
.globl  _restvr_26
_restvr_26:
        li      r12,-96
        lvx     v26,r12,r0
.globl  _restvr_27
_restvr_27:
        li      r12,-80
        lvx     v27,r12,r0
.globl  _restvr_28
_restvr_28:
        li      r12,-64
        lvx     v28,r12,r0
.globl  _restvr_29
_restvr_29:
        li      r12,-48
        lvx     v29,r12,r0
.globl  _restvr_30
_restvr_30:
        li      r12,-32
        lvx     v30,r12,r0
.globl  _restvr_31
_restvr_31:
        li      r12,-16
        lvx     v31,r12,r0
        blr

#endif /* CONFIG_ALTIVEC */

#endif /* CONFIG_PPC64 */

#endif