1 |
commit: 78d4a40a859636e46a150be8f53817faabb29744 |
2 |
Author: Michael Mair-Keimberger <m.mairkeimberger <AT> gmail <DOT> com> |
3 |
AuthorDate: Sat Apr 28 07:58:59 2018 +0000 |
4 |
Commit: Lars Wendler <polynomial-c <AT> gentoo <DOT> org> |
5 |
CommitDate: Fri May 4 07:18:28 2018 +0000 |
6 |
URL: https://gitweb.gentoo.org/repo/gentoo.git/commit/?id=78d4a40a |
7 |
|
8 |
dev-libs/openssl: remove unused patch |
9 |
|
10 |
Closes: https://github.com/gentoo/gentoo/pull/8189 |
11 |
|
12 |
.../files/openssl-1.1.0g-CVE-2017-3738.patch | 77 ---------------------- |
13 |
1 file changed, 77 deletions(-) |
14 |
|
15 |
diff --git a/dev-libs/openssl/files/openssl-1.1.0g-CVE-2017-3738.patch b/dev-libs/openssl/files/openssl-1.1.0g-CVE-2017-3738.patch |
16 |
deleted file mode 100644 |
17 |
index 4b01feb8e87..00000000000 |
18 |
--- a/dev-libs/openssl/files/openssl-1.1.0g-CVE-2017-3738.patch |
19 |
+++ /dev/null |
20 |
@@ -1,77 +0,0 @@ |
21 |
-From e502cc86df9dafded1694fceb3228ee34d11c11a Mon Sep 17 00:00:00 2001 |
22 |
-From: Andy Polyakov <appro@×××××××.org> |
23 |
-Date: Fri, 24 Nov 2017 11:35:50 +0100 |
24 |
-Subject: [PATCH] bn/asm/rsaz-avx2.pl: fix digit correction bug in |
25 |
- rsaz_1024_mul_avx2. |
26 |
- |
27 |
-Credit to OSS-Fuzz for finding this. |
28 |
- |
29 |
-CVE-2017-3738 |
30 |
- |
31 |
-Reviewed-by: Rich Salz <rsalz@×××××××.org> |
32 |
---- |
33 |
- crypto/bn/asm/rsaz-avx2.pl | 15 +++++++-------- |
34 |
- 1 file changed, 7 insertions(+), 8 deletions(-) |
35 |
- |
36 |
-diff --git a/crypto/bn/asm/rsaz-avx2.pl b/crypto/bn/asm/rsaz-avx2.pl |
37 |
-index 0c1b236ef98..46d746b7d0e 100755 |
38 |
---- a/crypto/bn/asm/rsaz-avx2.pl |
39 |
-+++ b/crypto/bn/asm/rsaz-avx2.pl |
40 |
-@@ -246,7 +246,7 @@ |
41 |
- vmovdqu 32*8-128($ap), $ACC8 |
42 |
- |
43 |
- lea 192(%rsp), $tp0 # 64+128=192 |
44 |
-- vpbroadcastq .Land_mask(%rip), $AND_MASK |
45 |
-+ vmovdqu .Land_mask(%rip), $AND_MASK |
46 |
- jmp .LOOP_GRANDE_SQR_1024 |
47 |
- |
48 |
- .align 32 |
49 |
-@@ -1077,10 +1077,10 @@ |
50 |
- vpmuludq 32*6-128($np),$Yi,$TEMP1 |
51 |
- vpaddq $TEMP1,$ACC6,$ACC6 |
52 |
- vpmuludq 32*7-128($np),$Yi,$TEMP2 |
53 |
-- vpblendd \$3, $ZERO, $ACC9, $ACC9 # correct $ACC3 |
54 |
-+ vpblendd \$3, $ZERO, $ACC9, $TEMP1 # correct $ACC3 |
55 |
- vpaddq $TEMP2,$ACC7,$ACC7 |
56 |
- vpmuludq 32*8-128($np),$Yi,$TEMP0 |
57 |
-- vpaddq $ACC9, $ACC3, $ACC3 # correct $ACC3 |
58 |
-+ vpaddq $TEMP1, $ACC3, $ACC3 # correct $ACC3 |
59 |
- vpaddq $TEMP0,$ACC8,$ACC8 |
60 |
- |
61 |
- mov %rbx, %rax |
62 |
-@@ -1093,7 +1093,9 @@ |
63 |
- vmovdqu -8+32*2-128($ap),$TEMP2 |
64 |
- |
65 |
- mov $r1, %rax |
66 |
-+ vpblendd \$0xfc, $ZERO, $ACC9, $ACC9 # correct $ACC3 |
67 |
- imull $n0, %eax |
68 |
-+ vpaddq $ACC9,$ACC4,$ACC4 # correct $ACC3 |
69 |
- and \$0x1fffffff, %eax |
70 |
- |
71 |
- imulq 16-128($ap),%rbx |
72 |
-@@ -1329,15 +1331,12 @@ |
73 |
- # But as we underutilize resources, it's possible to correct in |
74 |
- # each iteration with marginal performance loss. But then, as |
75 |
- # we do it in each iteration, we can correct less digits, and |
76 |
--# avoid performance penalties completely. Also note that we |
77 |
--# correct only three digits out of four. This works because |
78 |
--# most significant digit is subjected to less additions. |
79 |
-+# avoid performance penalties completely. |
80 |
- |
81 |
- $TEMP0 = $ACC9; |
82 |
- $TEMP3 = $Bi; |
83 |
- $TEMP4 = $Yi; |
84 |
- $code.=<<___; |
85 |
-- vpermq \$0, $AND_MASK, $AND_MASK |
86 |
- vpaddq (%rsp), $TEMP1, $ACC0 |
87 |
- |
88 |
- vpsrlq \$29, $ACC0, $TEMP1 |
89 |
-@@ -1770,7 +1769,7 @@ |
90 |
- |
91 |
- .align 64 |
92 |
- .Land_mask: |
93 |
-- .quad 0x1fffffff,0x1fffffff,0x1fffffff,-1 |
94 |
-+ .quad 0x1fffffff,0x1fffffff,0x1fffffff,0x1fffffff |
95 |
- .Lscatter_permd: |
96 |
- .long 0,2,4,6,7,7,7,7 |
97 |
- .Lgather_permd: |