1
0
mirror of https://github.com/hashcat/hashcat.git synced 2024-11-25 17:38:23 +00:00
hashcat/OpenCL/m15400_a3-optimized.cl
R. Yushaev 5de004103a Replace kernel parameter lists with macros
Substitute long parameter lists in ~2900 kernel function declarations
with macros. This cleans up the code, reduces probability of copy-paste
errors and highlights the differences between kernel functions. Also
reduces the size of the OpenCL folder by ~3 MB.
2018-11-16 11:44:33 +01:00

384 lines
7.5 KiB
Common Lisp

/**
* Author......: See docs/credits.txt
* License.....: MIT
*/
#define NEW_SIMD_CODE
#include "inc_vendor.cl"
#include "inc_hash_constants.h"
#include "inc_hash_functions.cl"
#include "inc_types.cl"
#include "inc_common.cl"
#include "inc_simd.cl"
#define CHACHA_CONST_00 0x61707865
#define CHACHA_CONST_01 0x3320646e
#define CHACHA_CONST_02 0x79622d32
#define CHACHA_CONST_03 0x6b206574
#define QR(a, b, c, d) \
do { \
x[a] = x[a] + x[b]; \
x[d] = rotl32(x[d] ^ x[a], 16); \
x[c] = x[c] + x[d]; \
x[b] = rotl32(x[b] ^ x[c], 12); \
x[a] = x[a] + x[b]; \
x[d] = rotl32(x[d] ^ x[a], 8); \
x[c] = x[c] + x[d]; \
x[b] = rotl32(x[b] ^ x[c], 7); \
} while (0);
DECLSPEC void chacha20_transform (const u32x *w0, const u32x *w1, const u32 *position, const u32 offset, const u32 *iv, const u32 *plain, u32x *digest)
{
/**
* Key expansion
*/
u32x ctx[16];
ctx[ 0] = CHACHA_CONST_00;
ctx[ 1] = CHACHA_CONST_01;
ctx[ 2] = CHACHA_CONST_02;
ctx[ 3] = CHACHA_CONST_03;
ctx[ 4] = w0[0];
ctx[ 5] = w0[1];
ctx[ 6] = w0[2];
ctx[ 7] = w0[3];
ctx[ 8] = w1[0];
ctx[ 9] = w1[1];
ctx[10] = w1[2];
ctx[11] = w1[3];
ctx[12] = position[0];
ctx[13] = position[1];
ctx[14] = iv[1];
ctx[15] = iv[0];
/**
* Generate 64 byte keystream
*/
u32x x[32];
x[ 0] = ctx[ 0];
x[ 1] = ctx[ 1];
x[ 2] = ctx[ 2];
x[ 3] = ctx[ 3];
x[ 4] = ctx[ 4];
x[ 5] = ctx[ 5];
x[ 6] = ctx[ 6];
x[ 7] = ctx[ 7];
x[ 8] = ctx[ 8];
x[ 9] = ctx[ 9];
x[10] = ctx[10];
x[11] = ctx[11];
x[12] = ctx[12];
x[13] = ctx[13];
x[14] = ctx[14];
x[15] = ctx[15];
#pragma unroll
for (u8 i = 0; i < 10; i++)
{
/* Column round */
QR(0, 4, 8, 12);
QR(1, 5, 9, 13);
QR(2, 6, 10, 14);
QR(3, 7, 11, 15);
/* Diagonal round */
QR(0, 5, 10, 15);
QR(1, 6, 11, 12);
QR(2, 7, 8, 13);
QR(3, 4, 9, 14);
}
x[ 0] += ctx[ 0];
x[ 1] += ctx[ 1];
x[ 2] += ctx[ 2];
x[ 3] += ctx[ 3];
x[ 4] += ctx[ 4];
x[ 5] += ctx[ 5];
x[ 6] += ctx[ 6];
x[ 7] += ctx[ 7];
x[ 8] += ctx[ 8];
x[ 9] += ctx[ 9];
x[10] += ctx[10];
x[11] += ctx[11];
x[12] += ctx[12];
x[13] += ctx[13];
x[14] += ctx[14];
x[15] += ctx[15];
if (offset > 56)
{
/**
* Generate a second 64 byte keystream
*/
ctx[12]++;
if (all(ctx[12] == 0)) ctx[13]++;
x[16] = ctx[ 0];
x[17] = ctx[ 1];
x[18] = ctx[ 2];
x[19] = ctx[ 3];
x[20] = ctx[ 4];
x[21] = ctx[ 5];
x[22] = ctx[ 6];
x[23] = ctx[ 7];
x[24] = ctx[ 8];
x[25] = ctx[ 9];
x[26] = ctx[10];
x[27] = ctx[11];
x[28] = ctx[12];
x[29] = ctx[13];
x[30] = ctx[14];
x[31] = ctx[15];
#pragma unroll
for (u8 i = 0; i < 10; i++)
{
/* Column round */
QR(16, 20, 24, 28);
QR(17, 21, 25, 29);
QR(18, 22, 26, 30);
QR(19, 23, 27, 31);
/* Diagonal round */
QR(16, 21, 26, 31);
QR(17, 22, 27, 28);
QR(18, 23, 24, 29);
QR(19, 20, 25, 30);
}
x[16] += ctx[ 0];
x[17] += ctx[ 1];
x[18] += ctx[ 2];
x[19] += ctx[ 3];
x[20] += ctx[ 4];
x[21] += ctx[ 5];
x[22] += ctx[ 6];
x[23] += ctx[ 7];
x[24] += ctx[ 8];
x[25] += ctx[ 9];
x[26] += ctx[10];
x[27] += ctx[11];
x[28] += ctx[12];
x[29] += ctx[13];
x[30] += ctx[14];
x[31] += ctx[15];
}
/**
* Encrypt plaintext with keystream
*/
const u32 index = offset / 4;
const u32 remain = offset % 4;
digest[0] = plain[1];
digest[1] = plain[0];
if (remain > 0)
{
digest[1] ^= x[index + 0] >> ( 0 + remain * 8);
digest[1] ^= x[index + 1] << (32 - remain * 8);
digest[0] ^= x[index + 1] >> ( 0 + remain * 8);
digest[0] ^= x[index + 2] << (32 - remain * 8);
}
else
{
digest[1] ^= x[index + 0];
digest[0] ^= x[index + 1];
}
}
__kernel void m15400_m04 (KERN_ATTR_VECTOR_ESALT (chacha20_t))
{
/**
* modifier
*/
const u64 gid = get_global_id (0);
const u64 lid = get_local_id (0);
u32 w0[4];
u32 w1[4];
w0[0] = pws[gid].i[0];
w0[1] = pws[gid].i[1];
w0[2] = pws[gid].i[2];
w0[3] = pws[gid].i[3];
w1[0] = pws[gid].i[4];
w1[1] = pws[gid].i[5];
w1[2] = pws[gid].i[6];
w1[3] = pws[gid].i[7];
u32x out_len = pws[gid].pw_len;
/**
* Salt prep
*/
u32 iv[2] = { 0 };
u32 plain[2] = { 0 };
u32 position[2] = { 0 };
u32 offset = 0;
position[0] = esalt_bufs[digests_offset].position[0];
position[1] = esalt_bufs[digests_offset].position[1];
offset = esalt_bufs[digests_offset].offset;
iv[0] = esalt_bufs[digests_offset].iv[0];
iv[1] = esalt_bufs[digests_offset].iv[1];
plain[0] = esalt_bufs[digests_offset].plain[0];
plain[1] = esalt_bufs[digests_offset].plain[1];
/**
* loop
*/
u32 w0l = pws[gid].i[0];
for (u32 il_pos = 0; il_pos < il_cnt; il_pos += VECT_SIZE)
{
const u32x w0r = words_buf_r[il_pos / VECT_SIZE];
const u32x w0x = w0l | w0r;
u32x w0_t[4];
u32x w1_t[4];
w0_t[0] = w0x;
w0_t[1] = w0[1];
w0_t[2] = w0[2];
w0_t[3] = w0[3];
w1_t[0] = w1[0];
w1_t[1] = w1[1];
w1_t[2] = w1[2];
w1_t[3] = w1[3];
u32x digest[4] = { 0 };
chacha20_transform (w0_t, w1_t, position, offset, iv, plain, digest);
const u32x r0 = digest[0];
const u32x r1 = digest[1];
const u32x r2 = digest[2];
const u32x r3 = digest[3];
COMPARE_M_SIMD(r0, r1, r2, r3);
}
}
__kernel void m15400_m08 (KERN_ATTR_ESALT (chacha20_t))
{
}
__kernel void m15400_m16 (KERN_ATTR_ESALT (chacha20_t))
{
}
__kernel void m15400_s04 (KERN_ATTR_VECTOR_ESALT (chacha20_t))
{
/**
* modifier
*/
const u64 gid = get_global_id (0);
const u64 lid = get_local_id (0);
u32 w0[4];
u32 w1[4];
w0[0] = pws[gid].i[0];
w0[1] = pws[gid].i[1];
w0[2] = pws[gid].i[2];
w0[3] = pws[gid].i[3];
w1[0] = pws[gid].i[4];
w1[1] = pws[gid].i[5];
w1[2] = pws[gid].i[6];
w1[3] = pws[gid].i[7];
u32 out_len = pws[gid].pw_len;
/**
* Salt prep
*/
u32 iv[2] = { 0 };
u32 plain[2] = { 0 };
u32 position[2] = { 0 };
u32 offset = 0;
position[0] = esalt_bufs[digests_offset].position[0];
position[1] = esalt_bufs[digests_offset].position[1];
offset = esalt_bufs[digests_offset].offset;
iv[0] = esalt_bufs[digests_offset].iv[0];
iv[1] = esalt_bufs[digests_offset].iv[1];
plain[0] = esalt_bufs[digests_offset].plain[0];
plain[1] = esalt_bufs[digests_offset].plain[1];
/**
* digest
*/
const u32 search[4] =
{
digests_buf[digests_offset].digest_buf[DGST_R0],
digests_buf[digests_offset].digest_buf[DGST_R1],
digests_buf[digests_offset].digest_buf[DGST_R2],
digests_buf[digests_offset].digest_buf[DGST_R3]
};
/**
* loop
*/
u32 w0l = pws[gid].i[0];
for (u32 il_pos = 0; il_pos < il_cnt; il_pos += VECT_SIZE)
{
const u32x w0r = words_buf_r[il_pos / VECT_SIZE];
const u32x w0x = w0l | w0r;
u32x w0_t[4];
u32x w1_t[4];
w0_t[0] = w0x;
w0_t[1] = w0[1];
w0_t[2] = w0[2];
w0_t[3] = w0[3];
w1_t[0] = w1[0];
w1_t[1] = w1[1];
w1_t[2] = w1[2];
w1_t[3] = w1[3];
u32x digest[4] = { 0 };
chacha20_transform (w0_t, w1_t, position, offset, iv, plain, digest);
const u32x r0 = digest[0];
const u32x r1 = digest[1];
const u32x r2 = digest[2];
const u32x r3 = digest[3];
COMPARE_S_SIMD(r0, r1, r2, r3);
}
}
__kernel void m15400_s08 (KERN_ATTR_ESALT (chacha20_t))
{
}
__kernel void m15400_s16 (KERN_ATTR_ESALT (chacha20_t))
{
}