From 2404dc8d0ac52ce5691a8782c12ad2ca784142ed Mon Sep 17 00:00:00 2001
From: =?UTF-8?q?Niels=20M=C3=B6ller?= <nisse@lysator.liu.se>
Date: Wed, 12 Oct 2005 22:56:03 +0200
Subject: [PATCH] New file, with first attempt at md5 assembler. Not yet
 working.

Rev: src/nettle/x86/md5-compress.asm:1.2
---
 x86/md5-compress.asm | 155 ++++++++++++++++++++++++++++++++++++++++++-
 1 file changed, 154 insertions(+), 1 deletion(-)

diff --git a/x86/md5-compress.asm b/x86/md5-compress.asm
index 234d35bf..47143bd2 100644
--- a/x86/md5-compress.asm
+++ b/x86/md5-compress.asm
@@ -22,8 +22,161 @@ define(<SA>,<%eax>)
 define(<SB>,<%ebx>)
 define(<SC>,<%ecx>)
 define(<SD>,<%edx>)
-
+define(<TMP>,<%ebp>)
+define(<TMP2>,<%edi>)
+define(<INPUT>,<%esi>)
 define(<DATA>,<%esp>)
 
 C %ebp, %edi and %esi available for temporaries
+
+C F1(x,y,z) = (z ^ x) & (y ^ z)
+define(<F1>, <
+	movl	$3, TMP2
+	movl	$3, TMP
+	xorl	$1, TMP2
+	xorl	$2, TMP
+	andl	TMP, TMP>)
+
+define(<F2>,<F1($3, $1, $2)>)
+
+C F3(x,y,z) = x ^ y ^ z
+define(<F3>,<
+	movl	$1, TMP
+	xorl	$2, TMP
+	xorl	$3, TMP>)
+
+C F4(x,y,z) = y ^ (x | ~z)
+define(<F4>,<
+	movl	$3, TMP
+	notl	TMP
+	orl	$1, TMP
+	xorl	$2, TMP>)
+
+define(<COPY>,<
+	movl	OFFSET($1)(INPUT), TMP
+	movl	TMP, OFFSET($1)(DATA)
+>)
+
+define(<REF>,<OFFSET(eval(4 * $1))(DATA)>)
+	
+C ROUND(f, w, x, y, z, k, data, s):
+C	w += f(x,y,z) + data + k
+C	w <<< s
+C	w += x
+define(<ROUND>,<
+	addl	$7, $2
+	$1($3, $4, $5)
+	addl	$6, $2
+	addl	TMP, $2
+	roll	<$>$8, $2
+	addl	$3, $2>)
+
+	.file "md5-compress.asm"
+
+	C _nettle_md5_compress(uint32_t *state, uint8_t *data)
 	
+	.text
+	.align 16
+PROLOGUE(_nettle_md5_compress)
+	C save all registers that need to be saved
+	
+	pushl	%ebx		C  76(%esp)
+	pushl	%ebp		C  72(%esp)
+	pushl	%esi		C  68(%esp)
+	pushl	%edi		C  64(%esp)
+
+	subl	$64, %esp	C  %esp = W
+
+	C load the state vector
+	movl	84(%esp),TMP
+	movl	(TMP),   SA
+	movl	4(TMP),  SB
+	movl	8(TMP),  SC
+	movl	12(TMP), SD
+
+	C Pointer to source data. We copy it to aligned storage
+	C at %esp when we first use it.
+	movl	88(%esp), INPUT
+
+	COPY( 0) ROUND(<F1>, SA, SB, SC, SD, TMP, $0xd76aa478, 7)
+	COPY( 1) ROUND(<F1>, SD, SA, SB, SC, TMP, $0xe8c7b756, 12)
+	COPY( 2) ROUND(<F1>, SC, SD, SA, SB, TMP, $0x242070db, 17)
+	COPY( 3) ROUND(<F1>, SB, SC, SD, SA, TMP, $0xc1bdceee, 22)
+	COPY( 4) ROUND(<F1>, SA, SB, SC, SD, TMP, $0xf57c0faf, 7)
+	COPY( 5) ROUND(<F1>, SD, SA, SB, SC, TMP, $0x4787c62a, 12)
+	COPY( 6) ROUND(<F1>, SC, SD, SA, SB, TMP, $0xa8304613, 17)
+	COPY( 7) ROUND(<F1>, SB, SC, SD, SA, TMP, $0xfd469501, 22)
+	COPY( 8) ROUND(<F1>, SA, SB, SC, SD, TMP, $0x698098d8, 7)
+	COPY( 9) ROUND(<F1>, SD, SA, SB, SC, TMP, $0x8b44f7af, 12)
+	COPY(10) ROUND(<F1>, SC, SD, SA, SB, TMP, $0xffff5bb1, 17)
+	COPY(11) ROUND(<F1>, SB, SC, SD, SA, TMP, $0x895cd7be, 22)
+	COPY(12) ROUND(<F1>, SA, SB, SC, SD, TMP, $0x6b901122, 7)
+	COPY(13) ROUND(<F1>, SD, SA, SB, SC, TMP, $0xfd987193, 12)
+	COPY(14) ROUND(<F1>, SC, SD, SA, SB, TMP, $0xa679438e, 17)
+	COPY(15) ROUND(<F1>, SB, SC, SD, SA, TMP, $0x49b40821, 22)
+	
+        ROUND(<F2>, SA, SB, SC, SD, REF( 1), $0xf61e2562, 5)
+        ROUND(<F2>, SD, SA, SB, SC, REF( 6), $0xc040b340, 9)
+        ROUND(<F2>, SC, SD, SA, SB, REF(11), $0x265e5a51, 14)
+        ROUND(<F2>, SB, SC, SD, SA, REF( 0), $0xe9b6c7aa, 20)
+        ROUND(<F2>, SA, SB, SC, SD, REF( 5), $0xd62f105d, 5)
+        ROUND(<F2>, SD, SA, SB, SC, REF(10), $0x02441453, 9)
+        ROUND(<F2>, SC, SD, SA, SB, REF(15), $0xd8a1e681, 14)
+        ROUND(<F2>, SB, SC, SD, SA, REF( 4), $0xe7d3fbc8, 20)
+        ROUND(<F2>, SA, SB, SC, SD, REF( 9), $0x21e1cde6, 5)
+        ROUND(<F2>, SD, SA, SB, SC, REF(14), $0xc33707d6, 9)
+        ROUND(<F2>, SC, SD, SA, SB, REF( 3), $0xf4d50d87, 14)
+        ROUND(<F2>, SB, SC, SD, SA, REF( 8), $0x455a14ed, 20)
+        ROUND(<F2>, SA, SB, SC, SD, REF(13), $0xa9e3e905, 5)
+        ROUND(<F2>, SD, SA, SB, SC, REF( 2), $0xfcefa3f8, 9)
+        ROUND(<F2>, SC, SD, SA, SB, REF( 7), $0x676f02d9, 14)
+        ROUND(<F2>, SB, SC, SD, SA, REF(12), $0x8d2a4c8a, 20)        
+
+        ROUND(<F3>, SA, SB, SC, SD, REF( 5), $0xfffa3942, 4)
+        ROUND(<F3>, SD, SA, SB, SC, REF( 8), $0x8771f681, 11)
+        ROUND(<F3>, SC, SD, SA, SB, REF(11), $0x6d9d6122, 16)
+        ROUND(<F3>, SB, SC, SD, SA, REF(14), $0xfde5380c, 23)
+        ROUND(<F3>, SA, SB, SC, SD, REF( 1), $0xa4beea44, 4)
+        ROUND(<F3>, SD, SA, SB, SC, REF( 4), $0x4bdecfa9, 11)
+        ROUND(<F3>, SC, SD, SA, SB, REF( 7), $0xf6bb4b60, 16)
+        ROUND(<F3>, SB, SC, SD, SA, REF(10), $0xbebfbc70, 23)
+        ROUND(<F3>, SA, SB, SC, SD, REF(13), $0x289b7ec6, 4)
+        ROUND(<F3>, SD, SA, SB, SC, REF( 0), $0xeaa127fa, 11)
+        ROUND(<F3>, SC, SD, SA, SB, REF( 3), $0xd4ef3085, 16)
+        ROUND(<F3>, SB, SC, SD, SA, REF( 6), $0x04881d05, 23)
+        ROUND(<F3>, SA, SB, SC, SD, REF( 9), $0xd9d4d039, 4)
+        ROUND(<F3>, SD, SA, SB, SC, REF(12), $0xe6db99e5, 11)
+        ROUND(<F3>, SC, SD, SA, SB, REF(15), $0x1fa27cf8, 16)
+        ROUND(<F3>, SB, SC, SD, SA, REF( 2), $0xc4ac5665, 23)        
+
+        ROUND(<F4>, SA, SB, SC, SD, REF( 0), $0xf4292244, 6)
+        ROUND(<F4>, SD, SA, SB, SC, REF( 7), $0x432aff97, 10)
+        ROUND(<F4>, SC, SD, SA, SB, REF(14), $0xab9423a7, 15)
+        ROUND(<F4>, SB, SC, SD, SA, REF( 5), $0xfc93a039, 21)
+        ROUND(<F4>, SA, SB, SC, SD, REF(12), $0x655b59c3, 6)
+        ROUND(<F4>, SD, SA, SB, SC, REF( 3), $0x8f0ccc92, 10)
+        ROUND(<F4>, SC, SD, SA, SB, REF(10), $0xffeff47d, 15)
+        ROUND(<F4>, SB, SC, SD, SA, REF( 1), $0x85845dd1, 21)
+        ROUND(<F4>, SA, SB, SC, SD, REF( 8), $0x6fa87e4f, 6)
+        ROUND(<F4>, SD, SA, SB, SC, REF(15), $0xfe2ce6e0, 10)
+        ROUND(<F4>, SC, SD, SA, SB, REF( 6), $0xa3014314, 15)
+        ROUND(<F4>, SB, SC, SD, SA, REF(13), $0x4e0811a1, 21)
+        ROUND(<F4>, SA, SB, SC, SD, REF( 4), $0xf7537e82, 6)
+        ROUND(<F4>, SD, SA, SB, SC, REF(11), $0xbd3af235, 10)
+        ROUND(<F4>, SC, SD, SA, SB, REF( 2), $0x2ad7d2bb, 15)
+        ROUND(<F4>, SB, SC, SD, SA, REF( 9), $0xeb86d391, 21)
+	
+	C Update the state vector
+	movl	84(%esp),TMP
+	addl	SA, (TMP) 
+	addl	SB, 4(TMP) 
+	addl	SC, 8(TMP) 
+	addl	SD, 12(TMP) 
+
+	addl	$64, %esp
+	popl	%edi
+	popl	%esi
+	popl	%ebp
+	popl	%ebx
+	ret
+EPILOGUE(_nettle_sha1_compress)
-- 
GitLab