aboutsummaryrefslogtreecommitdiff
path: root/vcpkg/ports/nettle/yasm.patch
diff options
context:
space:
mode:
Diffstat (limited to 'vcpkg/ports/nettle/yasm.patch')
-rw-r--r--vcpkg/ports/nettle/yasm.patch61
1 files changed, 61 insertions, 0 deletions
diff --git a/vcpkg/ports/nettle/yasm.patch b/vcpkg/ports/nettle/yasm.patch
new file mode 100644
index 0000000..5ea10ab
--- /dev/null
+++ b/vcpkg/ports/nettle/yasm.patch
@@ -0,0 +1,61 @@
+diff --git a/configure.ac b/configure.ac
+index e95f1c825..1d7ef36bb 100644
+--- a/configure.ac
++++ b/configure.ac
+@@ -302,7 +302,7 @@ ABI=standard
+ case "$host_cpu" in
+ [x86_64 | amd64])
+ AC_TRY_COMPILE([
+-#if defined(__x86_64__) || defined(__arch64__)
++#if defined(__x86_64__) || defined(__arch64__) || defined(_M_AMD64)
+ #error 64-bit x86
+ #endif
+ ], [], [
+diff --git a/x86_64/aes.m4 b/x86_64/aes.m4
+index 3bec9eae2..85cafc046 100644
+--- a/x86_64/aes.m4
++++ b/x86_64/aes.m4
+@@ -77,9 +77,9 @@ define(`AES_STORE', `
+ dnl AES_ROUND(table,a,b,c,d,out,ptr)
+ dnl Computes one word of the AES round. Leaves result in $6.
+ define(`AES_ROUND', `
+- movzb LREG($2), $7
++ movzbq LREG($2), $7
+ movl AES_TABLE0 ($1, $7, 4),$6
+- movzb HREG($3), XREG($7)
++ movzbq HREG($3), XREG($7)
+ xorl AES_TABLE1 ($1, $7, 4),$6
+ movl $4,XREG($7)
+ shr `$'16,$7
+@@ -94,7 +94,7 @@ dnl Computes one word of the final round. Leaves result in $6. Also
+ dnl performs the first substitution step, on the least significant
+ dnl byte, and rotates 8 bits.
+ define(`AES_FINAL_ROUND', `
+- movzb LREG($1),$7
++ movzbq LREG($1),$7
+ movzbl ($5, $7), $6
+ movl $2,XREG($7)
+ andl `$'0x0000ff00,XREG($7)
+@@ -113,18 +113,18 @@ dnl each of eax, ebx, ecx and edx, and also rotates
+ dnl the words one byte to the left.
+ dnl Uses that AES_SBOX == 0
+ define(`AES_SUBST_BYTE', `
+- movzb LREG($1),$6
++ movzbq LREG($1),$6
+ movb ($5, $6),LREG($1)
+ roll `$'8,$1
+
+- movzb LREG($2),$6
++ movzbq LREG($2),$6
+ movb ($5, $6),LREG($2)
+ roll `$'8,$2
+
+- movzb LREG($3),$6
++ movzbq LREG($3),$6
+ movb ($5, $6),LREG($3)
+ roll `$'8,$3
+
+- movzb LREG($4),$6
++ movzbq LREG($4),$6
+ movb ($5, $6),LREG($4)
+ roll `$'8,$4')dnl