c2ce6f9f3d
As our various loops (copy, string, crypto etc) get more complicated, we want to share implementations between userspace (eg glibc) and the kernel. We also want to write userspace test harnesses to put in tools/testing/selftest. One gratuitous difference between userspace and the kernel is the VMX register definitions - the kernel uses vrX whereas both gcc and glibc use vX. Change the kernel to match userspace. Signed-off-by: Anton Blanchard <anton@samba.org> Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
548 lines
11 KiB
ArmAsm
548 lines
11 KiB
ArmAsm
/*
|
|
* Special support for eabi and SVR4
|
|
*
|
|
* Copyright (C) 1995, 1996, 1998, 2000, 2001 Free Software Foundation, Inc.
|
|
* Copyright 2008 Freescale Semiconductor, Inc.
|
|
* Written By Michael Meissner
|
|
*
|
|
* Based on gcc/config/rs6000/crtsavres.asm from gcc
|
|
* 64 bit additions from reading the PPC elf64abi document.
|
|
*
|
|
* This file is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License as published by the
|
|
* Free Software Foundation; either version 2, or (at your option) any
|
|
* later version.
|
|
*
|
|
* In addition to the permissions in the GNU General Public License, the
|
|
* Free Software Foundation gives you unlimited permission to link the
|
|
* compiled version of this file with other programs, and to distribute
|
|
* those programs without any restriction coming from the use of this
|
|
* file. (The General Public License restrictions do apply in other
|
|
* respects; for example, they cover modification of the file, and
|
|
* distribution when not linked into another program.)
|
|
*
|
|
* This file is distributed in the hope that it will be useful, but
|
|
* WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; see the file COPYING. If not, write to
|
|
* the Free Software Foundation, 51 Franklin Street, Fifth Floor,
|
|
* Boston, MA 02110-1301, USA.
|
|
*
|
|
* As a special exception, if you link this library with files
|
|
* compiled with GCC to produce an executable, this does not cause
|
|
* the resulting executable to be covered by the GNU General Public License.
|
|
* This exception does not however invalidate any other reasons why
|
|
* the executable file might be covered by the GNU General Public License.
|
|
*/
|
|
|
|
#include <asm/ppc_asm.h>
|
|
|
|
.file "crtsavres.S"
|
|
|
|
#ifdef CONFIG_CC_OPTIMIZE_FOR_SIZE
|
|
|
|
#ifndef CONFIG_PPC64
|
|
|
|
.section ".text"
|
|
|
|
/* Routines for saving integer registers, called by the compiler. */
|
|
/* Called with r11 pointing to the stack header word of the caller of the */
|
|
/* function, just beyond the end of the integer save area. */
|
|
|
|
_GLOBAL(_savegpr_14)
|
|
_GLOBAL(_save32gpr_14)
|
|
stw 14,-72(11) /* save gp registers */
|
|
_GLOBAL(_savegpr_15)
|
|
_GLOBAL(_save32gpr_15)
|
|
stw 15,-68(11)
|
|
_GLOBAL(_savegpr_16)
|
|
_GLOBAL(_save32gpr_16)
|
|
stw 16,-64(11)
|
|
_GLOBAL(_savegpr_17)
|
|
_GLOBAL(_save32gpr_17)
|
|
stw 17,-60(11)
|
|
_GLOBAL(_savegpr_18)
|
|
_GLOBAL(_save32gpr_18)
|
|
stw 18,-56(11)
|
|
_GLOBAL(_savegpr_19)
|
|
_GLOBAL(_save32gpr_19)
|
|
stw 19,-52(11)
|
|
_GLOBAL(_savegpr_20)
|
|
_GLOBAL(_save32gpr_20)
|
|
stw 20,-48(11)
|
|
_GLOBAL(_savegpr_21)
|
|
_GLOBAL(_save32gpr_21)
|
|
stw 21,-44(11)
|
|
_GLOBAL(_savegpr_22)
|
|
_GLOBAL(_save32gpr_22)
|
|
stw 22,-40(11)
|
|
_GLOBAL(_savegpr_23)
|
|
_GLOBAL(_save32gpr_23)
|
|
stw 23,-36(11)
|
|
_GLOBAL(_savegpr_24)
|
|
_GLOBAL(_save32gpr_24)
|
|
stw 24,-32(11)
|
|
_GLOBAL(_savegpr_25)
|
|
_GLOBAL(_save32gpr_25)
|
|
stw 25,-28(11)
|
|
_GLOBAL(_savegpr_26)
|
|
_GLOBAL(_save32gpr_26)
|
|
stw 26,-24(11)
|
|
_GLOBAL(_savegpr_27)
|
|
_GLOBAL(_save32gpr_27)
|
|
stw 27,-20(11)
|
|
_GLOBAL(_savegpr_28)
|
|
_GLOBAL(_save32gpr_28)
|
|
stw 28,-16(11)
|
|
_GLOBAL(_savegpr_29)
|
|
_GLOBAL(_save32gpr_29)
|
|
stw 29,-12(11)
|
|
_GLOBAL(_savegpr_30)
|
|
_GLOBAL(_save32gpr_30)
|
|
stw 30,-8(11)
|
|
_GLOBAL(_savegpr_31)
|
|
_GLOBAL(_save32gpr_31)
|
|
stw 31,-4(11)
|
|
blr
|
|
|
|
/* Routines for restoring integer registers, called by the compiler. */
|
|
/* Called with r11 pointing to the stack header word of the caller of the */
|
|
/* function, just beyond the end of the integer restore area. */
|
|
|
|
_GLOBAL(_restgpr_14)
|
|
_GLOBAL(_rest32gpr_14)
|
|
lwz 14,-72(11) /* restore gp registers */
|
|
_GLOBAL(_restgpr_15)
|
|
_GLOBAL(_rest32gpr_15)
|
|
lwz 15,-68(11)
|
|
_GLOBAL(_restgpr_16)
|
|
_GLOBAL(_rest32gpr_16)
|
|
lwz 16,-64(11)
|
|
_GLOBAL(_restgpr_17)
|
|
_GLOBAL(_rest32gpr_17)
|
|
lwz 17,-60(11)
|
|
_GLOBAL(_restgpr_18)
|
|
_GLOBAL(_rest32gpr_18)
|
|
lwz 18,-56(11)
|
|
_GLOBAL(_restgpr_19)
|
|
_GLOBAL(_rest32gpr_19)
|
|
lwz 19,-52(11)
|
|
_GLOBAL(_restgpr_20)
|
|
_GLOBAL(_rest32gpr_20)
|
|
lwz 20,-48(11)
|
|
_GLOBAL(_restgpr_21)
|
|
_GLOBAL(_rest32gpr_21)
|
|
lwz 21,-44(11)
|
|
_GLOBAL(_restgpr_22)
|
|
_GLOBAL(_rest32gpr_22)
|
|
lwz 22,-40(11)
|
|
_GLOBAL(_restgpr_23)
|
|
_GLOBAL(_rest32gpr_23)
|
|
lwz 23,-36(11)
|
|
_GLOBAL(_restgpr_24)
|
|
_GLOBAL(_rest32gpr_24)
|
|
lwz 24,-32(11)
|
|
_GLOBAL(_restgpr_25)
|
|
_GLOBAL(_rest32gpr_25)
|
|
lwz 25,-28(11)
|
|
_GLOBAL(_restgpr_26)
|
|
_GLOBAL(_rest32gpr_26)
|
|
lwz 26,-24(11)
|
|
_GLOBAL(_restgpr_27)
|
|
_GLOBAL(_rest32gpr_27)
|
|
lwz 27,-20(11)
|
|
_GLOBAL(_restgpr_28)
|
|
_GLOBAL(_rest32gpr_28)
|
|
lwz 28,-16(11)
|
|
_GLOBAL(_restgpr_29)
|
|
_GLOBAL(_rest32gpr_29)
|
|
lwz 29,-12(11)
|
|
_GLOBAL(_restgpr_30)
|
|
_GLOBAL(_rest32gpr_30)
|
|
lwz 30,-8(11)
|
|
_GLOBAL(_restgpr_31)
|
|
_GLOBAL(_rest32gpr_31)
|
|
lwz 31,-4(11)
|
|
blr
|
|
|
|
/* Routines for restoring integer registers, called by the compiler. */
|
|
/* Called with r11 pointing to the stack header word of the caller of the */
|
|
/* function, just beyond the end of the integer restore area. */
|
|
|
|
_GLOBAL(_restgpr_14_x)
|
|
_GLOBAL(_rest32gpr_14_x)
|
|
lwz 14,-72(11) /* restore gp registers */
|
|
_GLOBAL(_restgpr_15_x)
|
|
_GLOBAL(_rest32gpr_15_x)
|
|
lwz 15,-68(11)
|
|
_GLOBAL(_restgpr_16_x)
|
|
_GLOBAL(_rest32gpr_16_x)
|
|
lwz 16,-64(11)
|
|
_GLOBAL(_restgpr_17_x)
|
|
_GLOBAL(_rest32gpr_17_x)
|
|
lwz 17,-60(11)
|
|
_GLOBAL(_restgpr_18_x)
|
|
_GLOBAL(_rest32gpr_18_x)
|
|
lwz 18,-56(11)
|
|
_GLOBAL(_restgpr_19_x)
|
|
_GLOBAL(_rest32gpr_19_x)
|
|
lwz 19,-52(11)
|
|
_GLOBAL(_restgpr_20_x)
|
|
_GLOBAL(_rest32gpr_20_x)
|
|
lwz 20,-48(11)
|
|
_GLOBAL(_restgpr_21_x)
|
|
_GLOBAL(_rest32gpr_21_x)
|
|
lwz 21,-44(11)
|
|
_GLOBAL(_restgpr_22_x)
|
|
_GLOBAL(_rest32gpr_22_x)
|
|
lwz 22,-40(11)
|
|
_GLOBAL(_restgpr_23_x)
|
|
_GLOBAL(_rest32gpr_23_x)
|
|
lwz 23,-36(11)
|
|
_GLOBAL(_restgpr_24_x)
|
|
_GLOBAL(_rest32gpr_24_x)
|
|
lwz 24,-32(11)
|
|
_GLOBAL(_restgpr_25_x)
|
|
_GLOBAL(_rest32gpr_25_x)
|
|
lwz 25,-28(11)
|
|
_GLOBAL(_restgpr_26_x)
|
|
_GLOBAL(_rest32gpr_26_x)
|
|
lwz 26,-24(11)
|
|
_GLOBAL(_restgpr_27_x)
|
|
_GLOBAL(_rest32gpr_27_x)
|
|
lwz 27,-20(11)
|
|
_GLOBAL(_restgpr_28_x)
|
|
_GLOBAL(_rest32gpr_28_x)
|
|
lwz 28,-16(11)
|
|
_GLOBAL(_restgpr_29_x)
|
|
_GLOBAL(_rest32gpr_29_x)
|
|
lwz 29,-12(11)
|
|
_GLOBAL(_restgpr_30_x)
|
|
_GLOBAL(_rest32gpr_30_x)
|
|
lwz 30,-8(11)
|
|
_GLOBAL(_restgpr_31_x)
|
|
_GLOBAL(_rest32gpr_31_x)
|
|
lwz 0,4(11)
|
|
lwz 31,-4(11)
|
|
mtlr 0
|
|
mr 1,11
|
|
blr
|
|
|
|
#ifdef CONFIG_ALTIVEC
|
|
/* Called with r0 pointing just beyond the end of the vector save area. */
|
|
|
|
_GLOBAL(_savevr_20)
|
|
li r11,-192
|
|
stvx v20,r11,r0
|
|
_GLOBAL(_savevr_21)
|
|
li r11,-176
|
|
stvx v21,r11,r0
|
|
_GLOBAL(_savevr_22)
|
|
li r11,-160
|
|
stvx v22,r11,r0
|
|
_GLOBAL(_savevr_23)
|
|
li r11,-144
|
|
stvx v23,r11,r0
|
|
_GLOBAL(_savevr_24)
|
|
li r11,-128
|
|
stvx v24,r11,r0
|
|
_GLOBAL(_savevr_25)
|
|
li r11,-112
|
|
stvx v25,r11,r0
|
|
_GLOBAL(_savevr_26)
|
|
li r11,-96
|
|
stvx v26,r11,r0
|
|
_GLOBAL(_savevr_27)
|
|
li r11,-80
|
|
stvx v27,r11,r0
|
|
_GLOBAL(_savevr_28)
|
|
li r11,-64
|
|
stvx v28,r11,r0
|
|
_GLOBAL(_savevr_29)
|
|
li r11,-48
|
|
stvx v29,r11,r0
|
|
_GLOBAL(_savevr_30)
|
|
li r11,-32
|
|
stvx v30,r11,r0
|
|
_GLOBAL(_savevr_31)
|
|
li r11,-16
|
|
stvx v31,r11,r0
|
|
blr
|
|
|
|
_GLOBAL(_restvr_20)
|
|
li r11,-192
|
|
lvx v20,r11,r0
|
|
_GLOBAL(_restvr_21)
|
|
li r11,-176
|
|
lvx v21,r11,r0
|
|
_GLOBAL(_restvr_22)
|
|
li r11,-160
|
|
lvx v22,r11,r0
|
|
_GLOBAL(_restvr_23)
|
|
li r11,-144
|
|
lvx v23,r11,r0
|
|
_GLOBAL(_restvr_24)
|
|
li r11,-128
|
|
lvx v24,r11,r0
|
|
_GLOBAL(_restvr_25)
|
|
li r11,-112
|
|
lvx v25,r11,r0
|
|
_GLOBAL(_restvr_26)
|
|
li r11,-96
|
|
lvx v26,r11,r0
|
|
_GLOBAL(_restvr_27)
|
|
li r11,-80
|
|
lvx v27,r11,r0
|
|
_GLOBAL(_restvr_28)
|
|
li r11,-64
|
|
lvx v28,r11,r0
|
|
_GLOBAL(_restvr_29)
|
|
li r11,-48
|
|
lvx v29,r11,r0
|
|
_GLOBAL(_restvr_30)
|
|
li r11,-32
|
|
lvx v30,r11,r0
|
|
_GLOBAL(_restvr_31)
|
|
li r11,-16
|
|
lvx v31,r11,r0
|
|
blr
|
|
|
|
#endif /* CONFIG_ALTIVEC */
|
|
|
|
#else /* CONFIG_PPC64 */
|
|
|
|
.section ".text.save.restore","ax",@progbits
|
|
|
|
.globl _savegpr0_14
|
|
_savegpr0_14:
|
|
std r14,-144(r1)
|
|
.globl _savegpr0_15
|
|
_savegpr0_15:
|
|
std r15,-136(r1)
|
|
.globl _savegpr0_16
|
|
_savegpr0_16:
|
|
std r16,-128(r1)
|
|
.globl _savegpr0_17
|
|
_savegpr0_17:
|
|
std r17,-120(r1)
|
|
.globl _savegpr0_18
|
|
_savegpr0_18:
|
|
std r18,-112(r1)
|
|
.globl _savegpr0_19
|
|
_savegpr0_19:
|
|
std r19,-104(r1)
|
|
.globl _savegpr0_20
|
|
_savegpr0_20:
|
|
std r20,-96(r1)
|
|
.globl _savegpr0_21
|
|
_savegpr0_21:
|
|
std r21,-88(r1)
|
|
.globl _savegpr0_22
|
|
_savegpr0_22:
|
|
std r22,-80(r1)
|
|
.globl _savegpr0_23
|
|
_savegpr0_23:
|
|
std r23,-72(r1)
|
|
.globl _savegpr0_24
|
|
_savegpr0_24:
|
|
std r24,-64(r1)
|
|
.globl _savegpr0_25
|
|
_savegpr0_25:
|
|
std r25,-56(r1)
|
|
.globl _savegpr0_26
|
|
_savegpr0_26:
|
|
std r26,-48(r1)
|
|
.globl _savegpr0_27
|
|
_savegpr0_27:
|
|
std r27,-40(r1)
|
|
.globl _savegpr0_28
|
|
_savegpr0_28:
|
|
std r28,-32(r1)
|
|
.globl _savegpr0_29
|
|
_savegpr0_29:
|
|
std r29,-24(r1)
|
|
.globl _savegpr0_30
|
|
_savegpr0_30:
|
|
std r30,-16(r1)
|
|
.globl _savegpr0_31
|
|
_savegpr0_31:
|
|
std r31,-8(r1)
|
|
std r0,16(r1)
|
|
blr
|
|
|
|
.globl _restgpr0_14
|
|
_restgpr0_14:
|
|
ld r14,-144(r1)
|
|
.globl _restgpr0_15
|
|
_restgpr0_15:
|
|
ld r15,-136(r1)
|
|
.globl _restgpr0_16
|
|
_restgpr0_16:
|
|
ld r16,-128(r1)
|
|
.globl _restgpr0_17
|
|
_restgpr0_17:
|
|
ld r17,-120(r1)
|
|
.globl _restgpr0_18
|
|
_restgpr0_18:
|
|
ld r18,-112(r1)
|
|
.globl _restgpr0_19
|
|
_restgpr0_19:
|
|
ld r19,-104(r1)
|
|
.globl _restgpr0_20
|
|
_restgpr0_20:
|
|
ld r20,-96(r1)
|
|
.globl _restgpr0_21
|
|
_restgpr0_21:
|
|
ld r21,-88(r1)
|
|
.globl _restgpr0_22
|
|
_restgpr0_22:
|
|
ld r22,-80(r1)
|
|
.globl _restgpr0_23
|
|
_restgpr0_23:
|
|
ld r23,-72(r1)
|
|
.globl _restgpr0_24
|
|
_restgpr0_24:
|
|
ld r24,-64(r1)
|
|
.globl _restgpr0_25
|
|
_restgpr0_25:
|
|
ld r25,-56(r1)
|
|
.globl _restgpr0_26
|
|
_restgpr0_26:
|
|
ld r26,-48(r1)
|
|
.globl _restgpr0_27
|
|
_restgpr0_27:
|
|
ld r27,-40(r1)
|
|
.globl _restgpr0_28
|
|
_restgpr0_28:
|
|
ld r28,-32(r1)
|
|
.globl _restgpr0_29
|
|
_restgpr0_29:
|
|
ld r0,16(r1)
|
|
ld r29,-24(r1)
|
|
mtlr r0
|
|
ld r30,-16(r1)
|
|
ld r31,-8(r1)
|
|
blr
|
|
|
|
.globl _restgpr0_30
|
|
_restgpr0_30:
|
|
ld r30,-16(r1)
|
|
.globl _restgpr0_31
|
|
_restgpr0_31:
|
|
ld r0,16(r1)
|
|
ld r31,-8(r1)
|
|
mtlr r0
|
|
blr
|
|
|
|
#ifdef CONFIG_ALTIVEC
|
|
/* Called with r0 pointing just beyond the end of the vector save area. */
|
|
|
|
.globl _savevr_20
|
|
_savevr_20:
|
|
li r12,-192
|
|
stvx v20,r12,r0
|
|
.globl _savevr_21
|
|
_savevr_21:
|
|
li r12,-176
|
|
stvx v21,r12,r0
|
|
.globl _savevr_22
|
|
_savevr_22:
|
|
li r12,-160
|
|
stvx v22,r12,r0
|
|
.globl _savevr_23
|
|
_savevr_23:
|
|
li r12,-144
|
|
stvx v23,r12,r0
|
|
.globl _savevr_24
|
|
_savevr_24:
|
|
li r12,-128
|
|
stvx v24,r12,r0
|
|
.globl _savevr_25
|
|
_savevr_25:
|
|
li r12,-112
|
|
stvx v25,r12,r0
|
|
.globl _savevr_26
|
|
_savevr_26:
|
|
li r12,-96
|
|
stvx v26,r12,r0
|
|
.globl _savevr_27
|
|
_savevr_27:
|
|
li r12,-80
|
|
stvx v27,r12,r0
|
|
.globl _savevr_28
|
|
_savevr_28:
|
|
li r12,-64
|
|
stvx v28,r12,r0
|
|
.globl _savevr_29
|
|
_savevr_29:
|
|
li r12,-48
|
|
stvx v29,r12,r0
|
|
.globl _savevr_30
|
|
_savevr_30:
|
|
li r12,-32
|
|
stvx v30,r12,r0
|
|
.globl _savevr_31
|
|
_savevr_31:
|
|
li r12,-16
|
|
stvx v31,r12,r0
|
|
blr
|
|
|
|
.globl _restvr_20
|
|
_restvr_20:
|
|
li r12,-192
|
|
lvx v20,r12,r0
|
|
.globl _restvr_21
|
|
_restvr_21:
|
|
li r12,-176
|
|
lvx v21,r12,r0
|
|
.globl _restvr_22
|
|
_restvr_22:
|
|
li r12,-160
|
|
lvx v22,r12,r0
|
|
.globl _restvr_23
|
|
_restvr_23:
|
|
li r12,-144
|
|
lvx v23,r12,r0
|
|
.globl _restvr_24
|
|
_restvr_24:
|
|
li r12,-128
|
|
lvx v24,r12,r0
|
|
.globl _restvr_25
|
|
_restvr_25:
|
|
li r12,-112
|
|
lvx v25,r12,r0
|
|
.globl _restvr_26
|
|
_restvr_26:
|
|
li r12,-96
|
|
lvx v26,r12,r0
|
|
.globl _restvr_27
|
|
_restvr_27:
|
|
li r12,-80
|
|
lvx v27,r12,r0
|
|
.globl _restvr_28
|
|
_restvr_28:
|
|
li r12,-64
|
|
lvx v28,r12,r0
|
|
.globl _restvr_29
|
|
_restvr_29:
|
|
li r12,-48
|
|
lvx v29,r12,r0
|
|
.globl _restvr_30
|
|
_restvr_30:
|
|
li r12,-32
|
|
lvx v30,r12,r0
|
|
.globl _restvr_31
|
|
_restvr_31:
|
|
li r12,-16
|
|
lvx v31,r12,r0
|
|
blr
|
|
|
|
#endif /* CONFIG_ALTIVEC */
|
|
|
|
#endif /* CONFIG_PPC64 */
|
|
|
|
#endif
|