123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397 |
- /*
- * test helper assembly functions
- *
- * Copyright (C) 2016 Simon Guo, IBM Corporation.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public License
- * as published by the Free Software Foundation; either version
- * 2 of the License, or (at your option) any later version.
- */
- #include <ppc-asm.h>
- #include "reg.h"
- /* Non volatile GPR - unsigned long buf[18] */
- FUNC_START(load_gpr)
- ld 14, 0*8(3)
- ld 15, 1*8(3)
- ld 16, 2*8(3)
- ld 17, 3*8(3)
- ld 18, 4*8(3)
- ld 19, 5*8(3)
- ld 20, 6*8(3)
- ld 21, 7*8(3)
- ld 22, 8*8(3)
- ld 23, 9*8(3)
- ld 24, 10*8(3)
- ld 25, 11*8(3)
- ld 26, 12*8(3)
- ld 27, 13*8(3)
- ld 28, 14*8(3)
- ld 29, 15*8(3)
- ld 30, 16*8(3)
- ld 31, 17*8(3)
- blr
- FUNC_END(load_gpr)
- FUNC_START(store_gpr)
- std 14, 0*8(3)
- std 15, 1*8(3)
- std 16, 2*8(3)
- std 17, 3*8(3)
- std 18, 4*8(3)
- std 19, 5*8(3)
- std 20, 6*8(3)
- std 21, 7*8(3)
- std 22, 8*8(3)
- std 23, 9*8(3)
- std 24, 10*8(3)
- std 25, 11*8(3)
- std 26, 12*8(3)
- std 27, 13*8(3)
- std 28, 14*8(3)
- std 29, 15*8(3)
- std 30, 16*8(3)
- std 31, 17*8(3)
- blr
- FUNC_END(store_gpr)
- /* Single Precision Float - float buf[32] */
- FUNC_START(load_fpr_single_precision)
- lfs 0, 0*4(3)
- lfs 1, 1*4(3)
- lfs 2, 2*4(3)
- lfs 3, 3*4(3)
- lfs 4, 4*4(3)
- lfs 5, 5*4(3)
- lfs 6, 6*4(3)
- lfs 7, 7*4(3)
- lfs 8, 8*4(3)
- lfs 9, 9*4(3)
- lfs 10, 10*4(3)
- lfs 11, 11*4(3)
- lfs 12, 12*4(3)
- lfs 13, 13*4(3)
- lfs 14, 14*4(3)
- lfs 15, 15*4(3)
- lfs 16, 16*4(3)
- lfs 17, 17*4(3)
- lfs 18, 18*4(3)
- lfs 19, 19*4(3)
- lfs 20, 20*4(3)
- lfs 21, 21*4(3)
- lfs 22, 22*4(3)
- lfs 23, 23*4(3)
- lfs 24, 24*4(3)
- lfs 25, 25*4(3)
- lfs 26, 26*4(3)
- lfs 27, 27*4(3)
- lfs 28, 28*4(3)
- lfs 29, 29*4(3)
- lfs 30, 30*4(3)
- lfs 31, 31*4(3)
- blr
- FUNC_END(load_fpr_single_precision)
- /* Single Precision Float - float buf[32] */
- FUNC_START(store_fpr_single_precision)
- stfs 0, 0*4(3)
- stfs 1, 1*4(3)
- stfs 2, 2*4(3)
- stfs 3, 3*4(3)
- stfs 4, 4*4(3)
- stfs 5, 5*4(3)
- stfs 6, 6*4(3)
- stfs 7, 7*4(3)
- stfs 8, 8*4(3)
- stfs 9, 9*4(3)
- stfs 10, 10*4(3)
- stfs 11, 11*4(3)
- stfs 12, 12*4(3)
- stfs 13, 13*4(3)
- stfs 14, 14*4(3)
- stfs 15, 15*4(3)
- stfs 16, 16*4(3)
- stfs 17, 17*4(3)
- stfs 18, 18*4(3)
- stfs 19, 19*4(3)
- stfs 20, 20*4(3)
- stfs 21, 21*4(3)
- stfs 22, 22*4(3)
- stfs 23, 23*4(3)
- stfs 24, 24*4(3)
- stfs 25, 25*4(3)
- stfs 26, 26*4(3)
- stfs 27, 27*4(3)
- stfs 28, 28*4(3)
- stfs 29, 29*4(3)
- stfs 30, 30*4(3)
- stfs 31, 31*4(3)
- blr
- FUNC_END(store_fpr_single_precision)
- /* VMX/VSX registers - unsigned long buf[128] */
- FUNC_START(loadvsx)
- lis 4, 0
- LXVD2X (0,(4),(3))
- addi 4, 4, 16
- LXVD2X (1,(4),(3))
- addi 4, 4, 16
- LXVD2X (2,(4),(3))
- addi 4, 4, 16
- LXVD2X (3,(4),(3))
- addi 4, 4, 16
- LXVD2X (4,(4),(3))
- addi 4, 4, 16
- LXVD2X (5,(4),(3))
- addi 4, 4, 16
- LXVD2X (6,(4),(3))
- addi 4, 4, 16
- LXVD2X (7,(4),(3))
- addi 4, 4, 16
- LXVD2X (8,(4),(3))
- addi 4, 4, 16
- LXVD2X (9,(4),(3))
- addi 4, 4, 16
- LXVD2X (10,(4),(3))
- addi 4, 4, 16
- LXVD2X (11,(4),(3))
- addi 4, 4, 16
- LXVD2X (12,(4),(3))
- addi 4, 4, 16
- LXVD2X (13,(4),(3))
- addi 4, 4, 16
- LXVD2X (14,(4),(3))
- addi 4, 4, 16
- LXVD2X (15,(4),(3))
- addi 4, 4, 16
- LXVD2X (16,(4),(3))
- addi 4, 4, 16
- LXVD2X (17,(4),(3))
- addi 4, 4, 16
- LXVD2X (18,(4),(3))
- addi 4, 4, 16
- LXVD2X (19,(4),(3))
- addi 4, 4, 16
- LXVD2X (20,(4),(3))
- addi 4, 4, 16
- LXVD2X (21,(4),(3))
- addi 4, 4, 16
- LXVD2X (22,(4),(3))
- addi 4, 4, 16
- LXVD2X (23,(4),(3))
- addi 4, 4, 16
- LXVD2X (24,(4),(3))
- addi 4, 4, 16
- LXVD2X (25,(4),(3))
- addi 4, 4, 16
- LXVD2X (26,(4),(3))
- addi 4, 4, 16
- LXVD2X (27,(4),(3))
- addi 4, 4, 16
- LXVD2X (28,(4),(3))
- addi 4, 4, 16
- LXVD2X (29,(4),(3))
- addi 4, 4, 16
- LXVD2X (30,(4),(3))
- addi 4, 4, 16
- LXVD2X (31,(4),(3))
- addi 4, 4, 16
- LXVD2X (32,(4),(3))
- addi 4, 4, 16
- LXVD2X (33,(4),(3))
- addi 4, 4, 16
- LXVD2X (34,(4),(3))
- addi 4, 4, 16
- LXVD2X (35,(4),(3))
- addi 4, 4, 16
- LXVD2X (36,(4),(3))
- addi 4, 4, 16
- LXVD2X (37,(4),(3))
- addi 4, 4, 16
- LXVD2X (38,(4),(3))
- addi 4, 4, 16
- LXVD2X (39,(4),(3))
- addi 4, 4, 16
- LXVD2X (40,(4),(3))
- addi 4, 4, 16
- LXVD2X (41,(4),(3))
- addi 4, 4, 16
- LXVD2X (42,(4),(3))
- addi 4, 4, 16
- LXVD2X (43,(4),(3))
- addi 4, 4, 16
- LXVD2X (44,(4),(3))
- addi 4, 4, 16
- LXVD2X (45,(4),(3))
- addi 4, 4, 16
- LXVD2X (46,(4),(3))
- addi 4, 4, 16
- LXVD2X (47,(4),(3))
- addi 4, 4, 16
- LXVD2X (48,(4),(3))
- addi 4, 4, 16
- LXVD2X (49,(4),(3))
- addi 4, 4, 16
- LXVD2X (50,(4),(3))
- addi 4, 4, 16
- LXVD2X (51,(4),(3))
- addi 4, 4, 16
- LXVD2X (52,(4),(3))
- addi 4, 4, 16
- LXVD2X (53,(4),(3))
- addi 4, 4, 16
- LXVD2X (54,(4),(3))
- addi 4, 4, 16
- LXVD2X (55,(4),(3))
- addi 4, 4, 16
- LXVD2X (56,(4),(3))
- addi 4, 4, 16
- LXVD2X (57,(4),(3))
- addi 4, 4, 16
- LXVD2X (58,(4),(3))
- addi 4, 4, 16
- LXVD2X (59,(4),(3))
- addi 4, 4, 16
- LXVD2X (60,(4),(3))
- addi 4, 4, 16
- LXVD2X (61,(4),(3))
- addi 4, 4, 16
- LXVD2X (62,(4),(3))
- addi 4, 4, 16
- LXVD2X (63,(4),(3))
- blr
- FUNC_END(loadvsx)
- FUNC_START(storevsx)
- lis 4, 0
- STXVD2X (0,(4),(3))
- addi 4, 4, 16
- STXVD2X (1,(4),(3))
- addi 4, 4, 16
- STXVD2X (2,(4),(3))
- addi 4, 4, 16
- STXVD2X (3,(4),(3))
- addi 4, 4, 16
- STXVD2X (4,(4),(3))
- addi 4, 4, 16
- STXVD2X (5,(4),(3))
- addi 4, 4, 16
- STXVD2X (6,(4),(3))
- addi 4, 4, 16
- STXVD2X (7,(4),(3))
- addi 4, 4, 16
- STXVD2X (8,(4),(3))
- addi 4, 4, 16
- STXVD2X (9,(4),(3))
- addi 4, 4, 16
- STXVD2X (10,(4),(3))
- addi 4, 4, 16
- STXVD2X (11,(4),(3))
- addi 4, 4, 16
- STXVD2X (12,(4),(3))
- addi 4, 4, 16
- STXVD2X (13,(4),(3))
- addi 4, 4, 16
- STXVD2X (14,(4),(3))
- addi 4, 4, 16
- STXVD2X (15,(4),(3))
- addi 4, 4, 16
- STXVD2X (16,(4),(3))
- addi 4, 4, 16
- STXVD2X (17,(4),(3))
- addi 4, 4, 16
- STXVD2X (18,(4),(3))
- addi 4, 4, 16
- STXVD2X (19,(4),(3))
- addi 4, 4, 16
- STXVD2X (20,(4),(3))
- addi 4, 4, 16
- STXVD2X (21,(4),(3))
- addi 4, 4, 16
- STXVD2X (22,(4),(3))
- addi 4, 4, 16
- STXVD2X (23,(4),(3))
- addi 4, 4, 16
- STXVD2X (24,(4),(3))
- addi 4, 4, 16
- STXVD2X (25,(4),(3))
- addi 4, 4, 16
- STXVD2X (26,(4),(3))
- addi 4, 4, 16
- STXVD2X (27,(4),(3))
- addi 4, 4, 16
- STXVD2X (28,(4),(3))
- addi 4, 4, 16
- STXVD2X (29,(4),(3))
- addi 4, 4, 16
- STXVD2X (30,(4),(3))
- addi 4, 4, 16
- STXVD2X (31,(4),(3))
- addi 4, 4, 16
- STXVD2X (32,(4),(3))
- addi 4, 4, 16
- STXVD2X (33,(4),(3))
- addi 4, 4, 16
- STXVD2X (34,(4),(3))
- addi 4, 4, 16
- STXVD2X (35,(4),(3))
- addi 4, 4, 16
- STXVD2X (36,(4),(3))
- addi 4, 4, 16
- STXVD2X (37,(4),(3))
- addi 4, 4, 16
- STXVD2X (38,(4),(3))
- addi 4, 4, 16
- STXVD2X (39,(4),(3))
- addi 4, 4, 16
- STXVD2X (40,(4),(3))
- addi 4, 4, 16
- STXVD2X (41,(4),(3))
- addi 4, 4, 16
- STXVD2X (42,(4),(3))
- addi 4, 4, 16
- STXVD2X (43,(4),(3))
- addi 4, 4, 16
- STXVD2X (44,(4),(3))
- addi 4, 4, 16
- STXVD2X (45,(4),(3))
- addi 4, 4, 16
- STXVD2X (46,(4),(3))
- addi 4, 4, 16
- STXVD2X (47,(4),(3))
- addi 4, 4, 16
- STXVD2X (48,(4),(3))
- addi 4, 4, 16
- STXVD2X (49,(4),(3))
- addi 4, 4, 16
- STXVD2X (50,(4),(3))
- addi 4, 4, 16
- STXVD2X (51,(4),(3))
- addi 4, 4, 16
- STXVD2X (52,(4),(3))
- addi 4, 4, 16
- STXVD2X (53,(4),(3))
- addi 4, 4, 16
- STXVD2X (54,(4),(3))
- addi 4, 4, 16
- STXVD2X (55,(4),(3))
- addi 4, 4, 16
- STXVD2X (56,(4),(3))
- addi 4, 4, 16
- STXVD2X (57,(4),(3))
- addi 4, 4, 16
- STXVD2X (58,(4),(3))
- addi 4, 4, 16
- STXVD2X (59,(4),(3))
- addi 4, 4, 16
- STXVD2X (60,(4),(3))
- addi 4, 4, 16
- STXVD2X (61,(4),(3))
- addi 4, 4, 16
- STXVD2X (62,(4),(3))
- addi 4, 4, 16
- STXVD2X (63,(4),(3))
- blr
- FUNC_END(storevsx)
|