diff --git a/powerpc64/fat/aes-decrypt-internal-2.asm b/powerpc64/fat/aes-decrypt-internal-2.asm index 3a4e08c2f7414c15cb6ca8072d73deed86a2523f..00db9fbf32f176fef1ec0e6f27cbb510ca28dcc8 100644 --- a/powerpc64/fat/aes-decrypt-internal-2.asm +++ b/powerpc64/fat/aes-decrypt-internal-2.asm @@ -1,7 +1,7 @@ C powerpc64/fat/aes-decrypt-internal-2.asm -ifelse(< +ifelse(` Copyright (C) 2020 Mamone Tarsha This file is part of GNU Nettle. @@ -29,9 +29,9 @@ ifelse(< You should have received copies of the GNU General Public License and the GNU Lesser General Public License along with this program. If not, see http://www.gnu.org/licenses/. ->) +') dnl PROLOGUE(_nettle_aes_decrypt) picked up by configure -define(<fat_transform>, <$1_ppc64>) -include_src(<powerpc64/p8/aes-decrypt-internal.asm>) +define(`fat_transform', `$1_ppc64') +include_src(`powerpc64/p8/aes-decrypt-internal.asm') diff --git a/powerpc64/fat/aes-encrypt-internal-2.asm b/powerpc64/fat/aes-encrypt-internal-2.asm index 42126e4f92e1b0d02fc819fd5c8a785465dbfecc..ded0aa74efa85b8270ce16c67627c2b0c121180c 100644 --- a/powerpc64/fat/aes-encrypt-internal-2.asm +++ b/powerpc64/fat/aes-encrypt-internal-2.asm @@ -1,7 +1,7 @@ C powerpc64/fat/aes-encrypt-internal-2.asm -ifelse(< +ifelse(` Copyright (C) 2020 Mamone Tarsha This file is part of GNU Nettle. @@ -29,9 +29,9 @@ ifelse(< You should have received copies of the GNU General Public License and the GNU Lesser General Public License along with this program. If not, see http://www.gnu.org/licenses/. ->) +') dnl PROLOGUE(_nettle_aes_encrypt) picked up by configure -define(<fat_transform>, <$1_ppc64>) -include_src(<powerpc64/p8/aes-encrypt-internal.asm>) +define(`fat_transform', `$1_ppc64') +include_src(`powerpc64/p8/aes-encrypt-internal.asm') diff --git a/powerpc64/machine.m4 b/powerpc64/machine.m4 index b76bb8b1dde9c7081ada1cf155b1c6ebd95de88e..221fa52334e97950945b5b9907d51d2ce59c8993 100644 --- a/powerpc64/machine.m4 +++ b/powerpc64/machine.m4 @@ -1,36 +1,36 @@ -define(<PROLOGUE>, -<.globl C_NAME($1) +define(`PROLOGUE', +`.globl C_NAME($1) DECLARE_FUNC(C_NAME($1)) ifelse(WORDS_BIGENDIAN,no, -<ifdef(<FUNC_ALIGN>,<.align FUNC_ALIGN>) +`ifdef(`FUNC_ALIGN',`.align FUNC_ALIGN') C_NAME($1): addis 2,12,(.TOC.-C_NAME($1))@ha addi 2,2,(.TOC.-C_NAME($1))@l -.localentry C_NAME($1), .-C_NAME($1)>, -<.section ".opd","aw" +.localentry C_NAME($1), .-C_NAME($1)', +`.section ".opd","aw" .align 3 C_NAME($1): .quad .C_NAME($1),.TOC.@tocbase,0 .previous -ifdef(<FUNC_ALIGN>,<.align FUNC_ALIGN>) -.C_NAME($1):>) -undefine(<FUNC_ALIGN>)>) +ifdef(`FUNC_ALIGN',`.align FUNC_ALIGN') +.C_NAME($1):') +undefine(`FUNC_ALIGN')') -define(<EPILOGUE>, -<ifelse(WORDS_BIGENDIAN,no, -<.size C_NAME($1), . - C_NAME($1)>, -<.size .C_NAME($1), . - .C_NAME($1) -.size C_NAME($1), . - .C_NAME($1)>)>) +define(`EPILOGUE', +`ifelse(WORDS_BIGENDIAN,no, +`.size C_NAME($1), . - C_NAME($1)', +`.size .C_NAME($1), . - .C_NAME($1) +.size C_NAME($1), . - .C_NAME($1)')') C Get vector-scalar register from vector register C VSR(VR) -define(<VSR>,<32+$1>) +define(`VSR',`32+$1') C Load the quadword in DATA_SRC storage into C VEC_DST. GPR is general-purpose register C used to obtain the effective address of C DATA_SRC storage. C DATA_LOAD_VEC(VEC_DST, DATA_SRC, GPR) -define(<DATA_LOAD_VEC>, -<ld $3,$2@got(2) -lvx $1,0,$3>) +define(`DATA_LOAD_VEC', +`ld $3,$2@got(2) +lvx $1,0,$3') diff --git a/powerpc64/p8/aes-decrypt-internal.asm b/powerpc64/p8/aes-decrypt-internal.asm index bfedb32b7d5bf00cd7d95880f20279e02e819a3a..acdbc1bd756282a1d28d539d1353b8771f09734d 100644 --- a/powerpc64/p8/aes-decrypt-internal.asm +++ b/powerpc64/p8/aes-decrypt-internal.asm @@ -1,6 +1,6 @@ C powerpc64/p8/aes-decrypt-internal.asm -ifelse(< +ifelse(` Copyright (C) 2020 Mamone Tarsha This file is part of GNU Nettle. @@ -27,36 +27,36 @@ ifelse(< You should have received copies of the GNU General Public License and the GNU Lesser General Public License along with this program. If not, see http://www.gnu.org/licenses/. ->) +') C Register usage: -define(<SP>, <1>) -define(<TOCP>, <2>) +define(`SP', `1') +define(`TOCP', `2') -define(<ROUNDS>, <3>) -define(<KEYS>, <4>) -define(<LENGTH>, <6>) -define(<DST>, <7>) -define(<SRC>, <8>) +define(`ROUNDS', `3') +define(`KEYS', `4') +define(`LENGTH', `6') +define(`DST', `7') +define(`SRC', `8') -define(<swap_mask>, <0>) +define(`swap_mask', `0') -define(<K>, <1>) -define(<S0>, <2>) -define(<S1>, <3>) -define(<S2>, <4>) -define(<S3>, <5>) -define(<S4>, <6>) -define(<S5>, <7>) -define(<S6>, <8>) -define(<S7>, <9>) +define(`K', `1') +define(`S0', `2') +define(`S1', `3') +define(`S2', `4') +define(`S3', `5') +define(`S4', `6') +define(`S5', `7') +define(`S6', `8') +define(`S7', `9') C ZERO vector register is used in place of RoundKey C for vncipher instruction because the order of InvMixColumns C and Xor processes are flipped in that instruction. C The Xor process with RoundKey is executed afterward. -define(<ZERO>, <10>) +define(`ZERO', `10') .file "aes-decrypt-internal.asm" @@ -67,7 +67,7 @@ define(<ZERO>, <10>) C size_t length, uint8_t *dst, C uint8_t *src) -define(<FUNC_ALIGN>, <5>) +define(`FUNC_ALIGN', `5') PROLOGUE(_nettle_aes_decrypt) vxor ZERO,ZERO,ZERO @@ -110,14 +110,14 @@ Lx8_loop: lxvd2x VSR(S6),30,SRC lxvd2x VSR(S7),31,SRC -IF_LE(<vperm S0,S0,S0,swap_mask +IF_LE(`vperm S0,S0,S0,swap_mask vperm S1,S1,S1,swap_mask vperm S2,S2,S2,swap_mask vperm S3,S3,S3,swap_mask vperm S4,S4,S4,swap_mask vperm S5,S5,S5,swap_mask vperm S6,S6,S6,swap_mask - vperm S7,S7,S7,swap_mask>) + vperm S7,S7,S7,swap_mask') vxor S0,S0,K vxor S1,S1,K @@ -164,14 +164,14 @@ L8x_round_loop: vncipherlast S6,S6,K vncipherlast S7,S7,K -IF_LE(<vperm S0,S0,S0,swap_mask +IF_LE(`vperm S0,S0,S0,swap_mask vperm S1,S1,S1,swap_mask vperm S2,S2,S2,swap_mask vperm S3,S3,S3,swap_mask vperm S4,S4,S4,swap_mask vperm S5,S5,S5,swap_mask vperm S6,S6,S6,swap_mask - vperm S7,S7,S7,swap_mask>) + vperm S7,S7,S7,swap_mask') stxvd2x VSR(S0),0,DST stxvd2x VSR(S1),25,DST @@ -213,10 +213,10 @@ L4x: addi 9,9,0x10 lxvd2x VSR(S3),9,SRC -IF_LE(<vperm S0,S0,S0,swap_mask +IF_LE(`vperm S0,S0,S0,swap_mask vperm S1,S1,S1,swap_mask vperm S2,S2,S2,swap_mask - vperm S3,S3,S3,swap_mask>) + vperm S3,S3,S3,swap_mask') vxor S0,S0,K vxor S1,S1,K @@ -247,10 +247,10 @@ L4x_round_loop: vncipherlast S2,S2,K vncipherlast S3,S3,K -IF_LE(<vperm S0,S0,S0,swap_mask +IF_LE(`vperm S0,S0,S0,swap_mask vperm S1,S1,S1,swap_mask vperm S2,S2,S2,swap_mask - vperm S3,S3,S3,swap_mask>) + vperm S3,S3,S3,swap_mask') stxvd2x VSR(S0),0,DST li 9,0x10 @@ -277,8 +277,8 @@ L2x: li 9,0x10 lxvd2x VSR(S1),9,SRC -IF_LE(<vperm S0,S0,S0,swap_mask - vperm S1,S1,S1,swap_mask>) +IF_LE(`vperm S0,S0,S0,swap_mask + vperm S1,S1,S1,swap_mask') vxor S0,S0,K vxor S1,S1,K @@ -301,8 +301,8 @@ L2x_round_loop: vncipherlast S0,S0,K vncipherlast S1,S1,K -IF_LE(<vperm S0,S0,S0,swap_mask - vperm S1,S1,S1,swap_mask>) +IF_LE(`vperm S0,S0,S0,swap_mask + vperm S1,S1,S1,swap_mask') stxvd2x VSR(S0),0,DST li 9,0x10 @@ -322,7 +322,7 @@ L1x: lxvd2x VSR(S0),0,SRC -IF_LE(<vperm S0,S0,S0,swap_mask>) +IF_LE(`vperm S0,S0,S0,swap_mask') vxor S0,S0,K @@ -341,7 +341,7 @@ L1x_round_loop: vperm K,K,K,swap_mask vncipherlast S0,S0,K -IF_LE(<vperm S0,S0,S0,swap_mask>) +IF_LE(`vperm S0,S0,S0,swap_mask') stxvd2x VSR(S0),0,DST @@ -352,5 +352,5 @@ EPILOGUE(_nettle_aes_decrypt) .data .align 4 .swap_mask: -IF_LE(<.byte 8,9,10,11,12,13,14,15,0,1,2,3,4,5,6,7>) -IF_BE(<.byte 3,2,1,0,7,6,5,4,11,10,9,8,15,14,13,12>) +IF_LE(`.byte 8,9,10,11,12,13,14,15,0,1,2,3,4,5,6,7') +IF_BE(`.byte 3,2,1,0,7,6,5,4,11,10,9,8,15,14,13,12') diff --git a/powerpc64/p8/aes-encrypt-internal.asm b/powerpc64/p8/aes-encrypt-internal.asm index 67c7e597807d62a05e86dcedd3af5991281f6066..482dff2567718b94b64d38cd5b2818f91e96c4fe 100644 --- a/powerpc64/p8/aes-encrypt-internal.asm +++ b/powerpc64/p8/aes-encrypt-internal.asm @@ -1,6 +1,6 @@ C powerpc64/p8/aes-encrypt-internal.asm -ifelse(< +ifelse(` Copyright (C) 2020 Mamone Tarsha This file is part of GNU Nettle. @@ -27,30 +27,30 @@ ifelse(< You should have received copies of the GNU General Public License and the GNU Lesser General Public License along with this program. If not, see http://www.gnu.org/licenses/. ->) +') C Register usage: -define(<SP>, <1>) -define(<TOCP>, <2>) +define(`SP', `1') +define(`TOCP', `2') -define(<ROUNDS>, <3>) -define(<KEYS>, <4>) -define(<LENGTH>, <6>) -define(<DST>, <7>) -define(<SRC>, <8>) +define(`ROUNDS', `3') +define(`KEYS', `4') +define(`LENGTH', `6') +define(`DST', `7') +define(`SRC', `8') -define(<swap_mask>, <0>) +define(`swap_mask', `0') -define(<K>, <1>) -define(<S0>, <2>) -define(<S1>, <3>) -define(<S2>, <4>) -define(<S3>, <5>) -define(<S4>, <6>) -define(<S5>, <7>) -define(<S6>, <8>) -define(<S7>, <9>) +define(`K', `1') +define(`S0', `2') +define(`S1', `3') +define(`S2', `4') +define(`S3', `5') +define(`S4', `6') +define(`S5', `7') +define(`S6', `8') +define(`S7', `9') .file "aes-encrypt-internal.asm" @@ -61,7 +61,7 @@ define(<S7>, <9>) C size_t length, uint8_t *dst, C uint8_t *src) -define(<FUNC_ALIGN>, <5>) +define(`FUNC_ALIGN', `5') PROLOGUE(_nettle_aes_encrypt) DATA_LOAD_VEC(swap_mask,.swap_mask,5) @@ -102,14 +102,14 @@ Lx8_loop: lxvd2x VSR(S6),30,SRC lxvd2x VSR(S7),31,SRC -IF_LE(<vperm S0,S0,S0,swap_mask +IF_LE(`vperm S0,S0,S0,swap_mask vperm S1,S1,S1,swap_mask vperm S2,S2,S2,swap_mask vperm S3,S3,S3,swap_mask vperm S4,S4,S4,swap_mask vperm S5,S5,S5,swap_mask vperm S6,S6,S6,swap_mask - vperm S7,S7,S7,swap_mask>) + vperm S7,S7,S7,swap_mask') vxor S0,S0,K vxor S1,S1,K @@ -148,14 +148,14 @@ L8x_round_loop: vcipherlast S6,S6,K vcipherlast S7,S7,K -IF_LE(<vperm S0,S0,S0,swap_mask +IF_LE(`vperm S0,S0,S0,swap_mask vperm S1,S1,S1,swap_mask vperm S2,S2,S2,swap_mask vperm S3,S3,S3,swap_mask vperm S4,S4,S4,swap_mask vperm S5,S5,S5,swap_mask vperm S6,S6,S6,swap_mask - vperm S7,S7,S7,swap_mask>) + vperm S7,S7,S7,swap_mask') stxvd2x VSR(S0),0,DST stxvd2x VSR(S1),25,DST @@ -197,10 +197,10 @@ L4x: addi 9,9,0x10 lxvd2x VSR(S3),9,SRC -IF_LE(<vperm S0,S0,S0,swap_mask +IF_LE(`vperm S0,S0,S0,swap_mask vperm S1,S1,S1,swap_mask vperm S2,S2,S2,swap_mask - vperm S3,S3,S3,swap_mask>) + vperm S3,S3,S3,swap_mask') vxor S0,S0,K vxor S1,S1,K @@ -227,10 +227,10 @@ L4x_round_loop: vcipherlast S2,S2,K vcipherlast S3,S3,K -IF_LE(<vperm S0,S0,S0,swap_mask +IF_LE(`vperm S0,S0,S0,swap_mask vperm S1,S1,S1,swap_mask vperm S2,S2,S2,swap_mask - vperm S3,S3,S3,swap_mask>) + vperm S3,S3,S3,swap_mask') stxvd2x VSR(S0),0,DST li 9,0x10 @@ -257,8 +257,8 @@ L2x: li 9,0x10 lxvd2x VSR(S1),9,SRC -IF_LE(<vperm S0,S0,S0,swap_mask - vperm S1,S1,S1,swap_mask>) +IF_LE(`vperm S0,S0,S0,swap_mask + vperm S1,S1,S1,swap_mask') vxor S0,S0,K vxor S1,S1,K @@ -279,8 +279,8 @@ L2x_round_loop: vcipherlast S0,S0,K vcipherlast S1,S1,K -IF_LE(<vperm S0,S0,S0,swap_mask - vperm S1,S1,S1,swap_mask>) +IF_LE(`vperm S0,S0,S0,swap_mask + vperm S1,S1,S1,swap_mask') stxvd2x VSR(S0),0,DST li 9,0x10 @@ -300,7 +300,7 @@ L1x: lxvd2x VSR(S0),0,SRC -IF_LE(<vperm S0,S0,S0,swap_mask>) +IF_LE(`vperm S0,S0,S0,swap_mask') vxor S0,S0,K @@ -318,7 +318,7 @@ L1x_round_loop: vperm K,K,K,swap_mask vcipherlast S0,S0,K -IF_LE(<vperm S0,S0,S0,swap_mask>) +IF_LE(`vperm S0,S0,S0,swap_mask') stxvd2x VSR(S0),0,DST @@ -329,5 +329,5 @@ EPILOGUE(_nettle_aes_encrypt) .data .align 4 .swap_mask: -IF_LE(<.byte 8,9,10,11,12,13,14,15,0,1,2,3,4,5,6,7>) -IF_BE(<.byte 3,2,1,0,7,6,5,4,11,10,9,8,15,14,13,12>) +IF_LE(`.byte 8,9,10,11,12,13,14,15,0,1,2,3,4,5,6,7') +IF_BE(`.byte 3,2,1,0,7,6,5,4,11,10,9,8,15,14,13,12')