398 lines
		
	
	
		
			7.0 KiB
		
	
	
	
		
			ArmAsm
		
	
	
	
	
	
			
		
		
	
	
			398 lines
		
	
	
		
			7.0 KiB
		
	
	
	
		
			ArmAsm
		
	
	
	
	
	
| /*
 | |
|  * test helper assembly functions
 | |
|  *
 | |
|  * Copyright (C) 2016 Simon Guo, IBM Corporation.
 | |
|  *
 | |
|  * This program is free software; you can redistribute it and/or
 | |
|  * modify it under the terms of the GNU General Public License
 | |
|  * as published by the Free Software Foundation; either version
 | |
|  * 2 of the License, or (at your option) any later version.
 | |
|  */
 | |
| #include <ppc-asm.h>
 | |
| #include "reg.h"
 | |
| 
 | |
| 
 | |
| /* Non volatile GPR - unsigned long buf[18] */
 | |
| FUNC_START(load_gpr)
 | |
| 	ld	14, 0*8(3)
 | |
| 	ld	15, 1*8(3)
 | |
| 	ld	16, 2*8(3)
 | |
| 	ld	17, 3*8(3)
 | |
| 	ld	18, 4*8(3)
 | |
| 	ld	19, 5*8(3)
 | |
| 	ld	20, 6*8(3)
 | |
| 	ld	21, 7*8(3)
 | |
| 	ld	22, 8*8(3)
 | |
| 	ld	23, 9*8(3)
 | |
| 	ld	24, 10*8(3)
 | |
| 	ld	25, 11*8(3)
 | |
| 	ld	26, 12*8(3)
 | |
| 	ld	27, 13*8(3)
 | |
| 	ld	28, 14*8(3)
 | |
| 	ld	29, 15*8(3)
 | |
| 	ld	30, 16*8(3)
 | |
| 	ld	31, 17*8(3)
 | |
| 	blr
 | |
| FUNC_END(load_gpr)
 | |
| 
 | |
| FUNC_START(store_gpr)
 | |
| 	std	14, 0*8(3)
 | |
| 	std	15, 1*8(3)
 | |
| 	std	16, 2*8(3)
 | |
| 	std	17, 3*8(3)
 | |
| 	std	18, 4*8(3)
 | |
| 	std	19, 5*8(3)
 | |
| 	std	20, 6*8(3)
 | |
| 	std	21, 7*8(3)
 | |
| 	std	22, 8*8(3)
 | |
| 	std	23, 9*8(3)
 | |
| 	std	24, 10*8(3)
 | |
| 	std	25, 11*8(3)
 | |
| 	std	26, 12*8(3)
 | |
| 	std	27, 13*8(3)
 | |
| 	std	28, 14*8(3)
 | |
| 	std	29, 15*8(3)
 | |
| 	std	30, 16*8(3)
 | |
| 	std	31, 17*8(3)
 | |
| 	blr
 | |
| FUNC_END(store_gpr)
 | |
| 
 | |
| /* Single Precision Float - float buf[32] */
 | |
| FUNC_START(load_fpr_single_precision)
 | |
| 	lfs 0, 0*4(3)
 | |
| 	lfs 1, 1*4(3)
 | |
| 	lfs 2, 2*4(3)
 | |
| 	lfs 3, 3*4(3)
 | |
| 	lfs 4, 4*4(3)
 | |
| 	lfs 5, 5*4(3)
 | |
| 	lfs 6, 6*4(3)
 | |
| 	lfs 7, 7*4(3)
 | |
| 	lfs 8, 8*4(3)
 | |
| 	lfs 9, 9*4(3)
 | |
| 	lfs 10, 10*4(3)
 | |
| 	lfs 11, 11*4(3)
 | |
| 	lfs 12, 12*4(3)
 | |
| 	lfs 13, 13*4(3)
 | |
| 	lfs 14, 14*4(3)
 | |
| 	lfs 15, 15*4(3)
 | |
| 	lfs 16, 16*4(3)
 | |
| 	lfs 17, 17*4(3)
 | |
| 	lfs 18, 18*4(3)
 | |
| 	lfs 19, 19*4(3)
 | |
| 	lfs 20, 20*4(3)
 | |
| 	lfs 21, 21*4(3)
 | |
| 	lfs 22, 22*4(3)
 | |
| 	lfs 23, 23*4(3)
 | |
| 	lfs 24, 24*4(3)
 | |
| 	lfs 25, 25*4(3)
 | |
| 	lfs 26, 26*4(3)
 | |
| 	lfs 27, 27*4(3)
 | |
| 	lfs 28, 28*4(3)
 | |
| 	lfs 29, 29*4(3)
 | |
| 	lfs 30, 30*4(3)
 | |
| 	lfs 31, 31*4(3)
 | |
| 	blr
 | |
| FUNC_END(load_fpr_single_precision)
 | |
| 
 | |
| /* Single Precision Float - float buf[32] */
 | |
| FUNC_START(store_fpr_single_precision)
 | |
| 	stfs 0, 0*4(3)
 | |
| 	stfs 1, 1*4(3)
 | |
| 	stfs 2, 2*4(3)
 | |
| 	stfs 3, 3*4(3)
 | |
| 	stfs 4, 4*4(3)
 | |
| 	stfs 5, 5*4(3)
 | |
| 	stfs 6, 6*4(3)
 | |
| 	stfs 7, 7*4(3)
 | |
| 	stfs 8, 8*4(3)
 | |
| 	stfs 9, 9*4(3)
 | |
| 	stfs 10, 10*4(3)
 | |
| 	stfs 11, 11*4(3)
 | |
| 	stfs 12, 12*4(3)
 | |
| 	stfs 13, 13*4(3)
 | |
| 	stfs 14, 14*4(3)
 | |
| 	stfs 15, 15*4(3)
 | |
| 	stfs 16, 16*4(3)
 | |
| 	stfs 17, 17*4(3)
 | |
| 	stfs 18, 18*4(3)
 | |
| 	stfs 19, 19*4(3)
 | |
| 	stfs 20, 20*4(3)
 | |
| 	stfs 21, 21*4(3)
 | |
| 	stfs 22, 22*4(3)
 | |
| 	stfs 23, 23*4(3)
 | |
| 	stfs 24, 24*4(3)
 | |
| 	stfs 25, 25*4(3)
 | |
| 	stfs 26, 26*4(3)
 | |
| 	stfs 27, 27*4(3)
 | |
| 	stfs 28, 28*4(3)
 | |
| 	stfs 29, 29*4(3)
 | |
| 	stfs 30, 30*4(3)
 | |
| 	stfs 31, 31*4(3)
 | |
| 	blr
 | |
| FUNC_END(store_fpr_single_precision)
 | |
| 
 | |
| /* VMX/VSX registers - unsigned long buf[128] */
 | |
| FUNC_START(loadvsx)
 | |
| 	lis	4, 0
 | |
| 	LXVD2X	(0,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(1,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(2,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(3,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(4,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(5,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(6,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(7,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(8,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(9,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(10,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(11,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(12,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(13,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(14,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(15,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(16,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(17,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(18,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(19,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(20,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(21,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(22,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(23,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(24,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(25,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(26,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(27,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(28,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(29,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(30,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(31,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(32,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(33,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(34,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(35,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(36,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(37,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(38,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(39,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(40,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(41,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(42,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(43,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(44,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(45,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(46,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(47,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(48,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(49,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(50,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(51,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(52,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(53,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(54,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(55,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(56,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(57,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(58,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(59,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(60,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(61,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(62,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	LXVD2X	(63,(4),(3))
 | |
| 	blr
 | |
| FUNC_END(loadvsx)
 | |
| 
 | |
| FUNC_START(storevsx)
 | |
| 	lis	4, 0
 | |
| 	STXVD2X	(0,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(1,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(2,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(3,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(4,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(5,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(6,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(7,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(8,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(9,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(10,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(11,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(12,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(13,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(14,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(15,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(16,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(17,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(18,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(19,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(20,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(21,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(22,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(23,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(24,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(25,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(26,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(27,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(28,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(29,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(30,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(31,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(32,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(33,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(34,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(35,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(36,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(37,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(38,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(39,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(40,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(41,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(42,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(43,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(44,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(45,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(46,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(47,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(48,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(49,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(50,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(51,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(52,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(53,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(54,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(55,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(56,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(57,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(58,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(59,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(60,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(61,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(62,(4),(3))
 | |
| 	addi	4, 4, 16
 | |
| 	STXVD2X	(63,(4),(3))
 | |
| 	blr
 | |
| FUNC_END(storevsx)
 | 
