nixpkgs/pkgs/tools/archivers/7zz/fix-empty-expr-stmt.patch
2023-08-22 21:19:21 -04:00

185 lines
6.4 KiB
Diff

diff -Naur a/C/7zCrc.c b/C/7zCrc.c
--- a/C/7zCrc.c
+++ b/C/7zCrc.c
@@ -174,7 +174,7 @@
UInt32 Z7_FASTCALL CrcUpdateT0_32(UInt32 v, const void *data, size_t size, const UInt32 *table)
{
const Byte *p = (const Byte *)data;
- UNUSED_VAR(table);
+ UNUSED_VAR(table)
for (; size != 0 && ((unsigned)(ptrdiff_t)p & (T0_32_UNROLL_BYTES - 1)) != 0; size--)
v = __crc32b(v, *p++);
@@ -206,7 +206,7 @@
UInt32 Z7_FASTCALL CrcUpdateT0_64(UInt32 v, const void *data, size_t size, const UInt32 *table)
{
const Byte *p = (const Byte *)data;
- UNUSED_VAR(table);
+ UNUSED_VAR(table)
for (; size != 0 && ((unsigned)(ptrdiff_t)p & (T0_64_UNROLL_BYTES - 1)) != 0; size--)
v = __crc32b(v, *p++);
diff -Naur a/C/AesOpt.c b/C/AesOpt.c
--- a/C/AesOpt.c
+++ b/C/AesOpt.c
@@ -639,7 +639,7 @@
const v128 k_z0 = w[2];
for (; numBlocks != 0; numBlocks--, data++)
{
- MM_XOR_m (*data);
+ MM_XOR_m (*data)
AES_E_MC_m (k0)
AES_E_MC_m (k1)
AES_E_MC_m (k2)
@@ -660,7 +660,7 @@
}
}
AES_E_m (k_z1)
- MM_XOR_m (k_z0);
+ MM_XOR_m (k_z0)
*data = m;
}
*p = m;
@@ -745,7 +745,7 @@
while (w != p);
WOP_KEY (AES_D, 1)
WOP_KEY (AES_XOR, 0)
- MM_XOR (m0, iv);
+ MM_XOR (m0, iv)
WOP_M1 (XOR_data_M1)
iv = data[NUM_WAYS - 1];
WOP (STORE_data)
@@ -759,14 +759,14 @@
AES_D_IMC_m (w[2])
do
{
- AES_D_IMC_m (w[1]);
- AES_D_IMC_m (w[0]);
+ AES_D_IMC_m (w[1])
+ AES_D_IMC_m (w[0])
w -= 2;
}
while (w != p);
- AES_D_m (w[1]);
- MM_XOR_m (w[0]);
- MM_XOR_m (iv);
+ AES_D_m (w[1])
+ MM_XOR_m (w[0])
+ MM_XOR_m (iv)
iv = *data;
*data = m;
}
@@ -809,11 +809,11 @@
{
const v128 *w = p;
v128 m;
- CTR_START (m, 0);
+ CTR_START (m, 0)
do
{
- AES_E_MC_m (w[0]);
- AES_E_MC_m (w[1]);
+ AES_E_MC_m (w[0])
+ AES_E_MC_m (w[1])
w += 2;
}
while (w != wEnd);
diff -Naur a/C/LzmaEnc.c b/C/LzmaEnc.c
--- a/C/LzmaEnc.c
+++ b/C/LzmaEnc.c
@@ -195,11 +195,11 @@
unsigned GetPosSlot1(UInt32 pos)
{
unsigned res;
- BSR2_RET(pos, res);
+ BSR2_RET(pos, res)
return res;
}
-#define GetPosSlot2(pos, res) { BSR2_RET(pos, res); }
-#define GetPosSlot(pos, res) { if (pos < 2) res = pos; else BSR2_RET(pos, res); }
+#define GetPosSlot2(pos, res) { BSR2_RET(pos, res) }
+#define GetPosSlot(pos, res) { if (pos < 2) res = pos; else BSR2_RET(pos, res) }
#else // ! LZMA_LOG_BSR
diff -Naur a/C/Sha1Opt.c b/C/Sha1Opt.c
--- a/C/Sha1Opt.c
+++ b/C/Sha1Opt.c
@@ -300,26 +300,26 @@
LOAD_SHUFFLE (m2, 2)
LOAD_SHUFFLE (m3, 3)
- T(m0, c0); H(e1); C(e0);
- T(m1, c0); SU0(m0, m1, m2); H(e0); C(e1);
- T(m2, c0); SU0(m1, m2, m3); SU1(m0, m3); H(e1); C(e0);
- T(m3, c0); SU0(m2, m3, m0); SU1(m1, m0); H(e0); C(e1);
- T(m0, c0); SU0(m3, m0, m1); SU1(m2, m1); H(e1); C(e0);
- T(m1, c1); SU0(m0, m1, m2); SU1(m3, m2); H(e0); P(e1);
- T(m2, c1); SU0(m1, m2, m3); SU1(m0, m3); H(e1); P(e0);
- T(m3, c1); SU0(m2, m3, m0); SU1(m1, m0); H(e0); P(e1);
- T(m0, c1); SU0(m3, m0, m1); SU1(m2, m1); H(e1); P(e0);
- T(m1, c1); SU0(m0, m1, m2); SU1(m3, m2); H(e0); P(e1);
- T(m2, c2); SU0(m1, m2, m3); SU1(m0, m3); H(e1); M(e0);
- T(m3, c2); SU0(m2, m3, m0); SU1(m1, m0); H(e0); M(e1);
- T(m0, c2); SU0(m3, m0, m1); SU1(m2, m1); H(e1); M(e0);
- T(m1, c2); SU0(m0, m1, m2); SU1(m3, m2); H(e0); M(e1);
- T(m2, c2); SU0(m1, m2, m3); SU1(m0, m3); H(e1); M(e0);
- T(m3, c3); SU0(m2, m3, m0); SU1(m1, m0); H(e0); P(e1);
- T(m0, c3); SU0(m3, m0, m1); SU1(m2, m1); H(e1); P(e0);
- T(m1, c3); SU1(m3, m2); H(e0); P(e1);
- T(m2, c3); H(e1); P(e0);
- T(m3, c3); H(e0); P(e1);
+ T(m0, c0); H(e1); C(e0)
+ T(m1, c0); SU0(m0, m1, m2) H(e0); C(e1)
+ T(m2, c0); SU0(m1, m2, m3) SU1(m0, m3) H(e1); C(e0)
+ T(m3, c0); SU0(m2, m3, m0) SU1(m1, m0) H(e0); C(e1)
+ T(m0, c0); SU0(m3, m0, m1) SU1(m2, m1) H(e1); C(e0)
+ T(m1, c1); SU0(m0, m1, m2) SU1(m3, m2) H(e0); P(e1)
+ T(m2, c1); SU0(m1, m2, m3) SU1(m0, m3) H(e1); P(e0)
+ T(m3, c1); SU0(m2, m3, m0) SU1(m1, m0) H(e0); P(e1)
+ T(m0, c1); SU0(m3, m0, m1) SU1(m2, m1) H(e1); P(e0)
+ T(m1, c1); SU0(m0, m1, m2) SU1(m3, m2) H(e0); P(e1)
+ T(m2, c2); SU0(m1, m2, m3) SU1(m0, m3) H(e1); M(e0)
+ T(m3, c2); SU0(m2, m3, m0) SU1(m1, m0) H(e0); M(e1)
+ T(m0, c2); SU0(m3, m0, m1) SU1(m2, m1) H(e1); M(e0)
+ T(m1, c2); SU0(m0, m1, m2) SU1(m3, m2) H(e0); M(e1)
+ T(m2, c2); SU0(m1, m2, m3) SU1(m0, m3) H(e1); M(e0)
+ T(m3, c3); SU0(m2, m3, m0) SU1(m1, m0) H(e0); P(e1)
+ T(m0, c3); SU0(m3, m0, m1) SU1(m2, m1) H(e1); P(e0)
+ T(m1, c3); SU1(m3, m2) H(e0); P(e1)
+ T(m2, c3); H(e1); P(e0)
+ T(m3, c3); H(e0); P(e1)
abcd = vaddq_u32(abcd, abcd_save);
e0 += e0_save;
diff -Naur a/C/Sha256Opt.c b/C/Sha256Opt.c
--- a/C/Sha256Opt.c
+++ b/C/Sha256Opt.c
@@ -316,10 +316,10 @@
LOAD_SHUFFLE (m2, 2)
LOAD_SHUFFLE (m3, 3)
- R16 ( 0, NNN, NNN, SM1, NNN, SM1, SM2, SM1, SM2 );
- R16 ( 1, SM1, SM2, SM1, SM2, SM1, SM2, SM1, SM2 );
- R16 ( 2, SM1, SM2, SM1, SM2, SM1, SM2, SM1, SM2 );
- R16 ( 3, SM1, SM2, NNN, SM2, NNN, NNN, NNN, NNN );
+ R16 ( 0, NNN, NNN, SM1, NNN, SM1, SM2, SM1, SM2 )
+ R16 ( 1, SM1, SM2, SM1, SM2, SM1, SM2, SM1, SM2 )
+ R16 ( 2, SM1, SM2, SM1, SM2, SM1, SM2, SM1, SM2 )
+ R16 ( 3, SM1, SM2, NNN, SM2, NNN, NNN, NNN, NNN )
state0 = vaddq_u32(state0, state0_save);
state1 = vaddq_u32(state1, state1_save);
diff -Naur a/CPP/Windows/System.h b/CPP/Windows/System.h
--- a/CPP/Windows/System.h
+++ b/CPP/Windows/System.h
@@ -105,7 +105,7 @@
}
void CpuZero() { }
- void CpuSet(unsigned cpuIndex) { UNUSED_VAR(cpuIndex); }
+ void CpuSet(unsigned cpuIndex) { UNUSED_VAR(cpuIndex) }
int IsCpuSet(unsigned cpuIndex) const { return (cpuIndex < numSysThreads) ? 1 : 0; }
BOOL SetProcAffinity() const