mirror of
https://github.com/hashcat/hashcat.git
synced 2025-07-06 06:42:35 +00:00

Updated kernel declarations from "KERNEL_FQ void HC_ATTR_SEQ" to "KERNEL_FQ KERNEL_FA void". Please update your custom plugin kernels accordingly. Added spilling size as a factor in calculating usable memory per device. This is based on undocumented variables and may not be 100% accurate, but it works well in practice. Added a compiler hint to scrypt-based kernels indicating the guaranteed maximum thread count per kernel invocation. Removed redundant kernel code 29800, as it is identical to 27700, and updated the plugin.
417 lines
7.9 KiB
Common Lisp
417 lines
7.9 KiB
Common Lisp
/**
|
|
* Author......: See docs/credits.txt
|
|
* License.....: MIT
|
|
*/
|
|
|
|
//too much register pressure
|
|
//#define NEW_SIMD_CODE
|
|
|
|
#ifdef KERNEL_STATIC
|
|
#include M2S(INCLUDE_PATH/inc_vendor.h)
|
|
#include M2S(INCLUDE_PATH/inc_types.h)
|
|
#include M2S(INCLUDE_PATH/inc_platform.cl)
|
|
#include M2S(INCLUDE_PATH/inc_common.cl)
|
|
#include M2S(INCLUDE_PATH/inc_rp_optimized.h)
|
|
#include M2S(INCLUDE_PATH/inc_rp_optimized.cl)
|
|
#include M2S(INCLUDE_PATH/inc_simd.cl)
|
|
#include M2S(INCLUDE_PATH/inc_hash_streebog256.cl)
|
|
#endif
|
|
|
|
#define INITVAL 0x0101010101010101UL
|
|
|
|
DECLSPEC void streebog_g (PRIVATE_AS u64x *h, PRIVATE_AS const u64x *m, LOCAL_AS u64 (*s_sbob_sl64)[256])
|
|
{
|
|
u64x k[8];
|
|
u64x s[8];
|
|
u64x t[8];
|
|
|
|
#ifdef _unroll
|
|
#pragma unroll
|
|
#endif
|
|
for (int i = 0; i < 8; i++)
|
|
{
|
|
t[i] = h[i];
|
|
}
|
|
|
|
for (int i = 0; i < 8; i++)
|
|
{
|
|
k[i] = SBOG_LPSti64;
|
|
}
|
|
|
|
#ifdef _unroll
|
|
#pragma unroll
|
|
#endif
|
|
for (int i = 0; i < 8; i++)
|
|
{
|
|
s[i] = m[i];
|
|
}
|
|
|
|
for (int r = 0; r < 12; r++)
|
|
{
|
|
#ifdef _unroll
|
|
#pragma unroll
|
|
#endif
|
|
for (int i = 0; i < 8; i++)
|
|
{
|
|
t[i] = s[i] ^ k[i];
|
|
}
|
|
|
|
#ifdef _unroll
|
|
#pragma unroll
|
|
#endif
|
|
for (int i = 0; i < 8; i++)
|
|
{
|
|
s[i] = SBOG_LPSti64;
|
|
}
|
|
|
|
for (int i = 0; i < 8; i++)
|
|
{
|
|
t[i] = k[i] ^ sbob256_rc64[r][i];
|
|
}
|
|
|
|
#ifdef _unroll
|
|
#pragma unroll
|
|
#endif
|
|
for (int i = 0; i < 8; i++)
|
|
{
|
|
k[i] = SBOG_LPSti64;
|
|
}
|
|
}
|
|
|
|
#ifdef _unroll
|
|
#pragma unroll
|
|
#endif
|
|
for (int i = 0; i < 8; i++)
|
|
{
|
|
h[i] ^= s[i] ^ k[i] ^ m[i];
|
|
}
|
|
}
|
|
|
|
KERNEL_FQ KERNEL_FA void m11700_m04 (KERN_ATTR_RULES ())
|
|
{
|
|
/**
|
|
* modifier
|
|
*/
|
|
|
|
const u64 gid = get_global_id (0);
|
|
const u64 lid = get_local_id (0);
|
|
const u64 lsz = get_local_size (0);
|
|
|
|
/**
|
|
* shared lookup table
|
|
*/
|
|
|
|
LOCAL_VK u64 s_sbob_sl64[8][256];
|
|
|
|
for (u32 i = lid; i < 256; i += lsz)
|
|
{
|
|
s_sbob_sl64[0][i] = sbob256_sl64[0][i];
|
|
s_sbob_sl64[1][i] = sbob256_sl64[1][i];
|
|
s_sbob_sl64[2][i] = sbob256_sl64[2][i];
|
|
s_sbob_sl64[3][i] = sbob256_sl64[3][i];
|
|
s_sbob_sl64[4][i] = sbob256_sl64[4][i];
|
|
s_sbob_sl64[5][i] = sbob256_sl64[5][i];
|
|
s_sbob_sl64[6][i] = sbob256_sl64[6][i];
|
|
s_sbob_sl64[7][i] = sbob256_sl64[7][i];
|
|
}
|
|
|
|
SYNC_THREADS ();
|
|
|
|
if (gid >= GID_CNT) return;
|
|
|
|
/**
|
|
* base
|
|
*/
|
|
|
|
u32 pw_buf0[4];
|
|
u32 pw_buf1[4];
|
|
|
|
pw_buf0[0] = pws[gid].i[ 0];
|
|
pw_buf0[1] = pws[gid].i[ 1];
|
|
pw_buf0[2] = pws[gid].i[ 2];
|
|
pw_buf0[3] = pws[gid].i[ 3];
|
|
pw_buf1[0] = pws[gid].i[ 4];
|
|
pw_buf1[1] = pws[gid].i[ 5];
|
|
pw_buf1[2] = pws[gid].i[ 6];
|
|
pw_buf1[3] = pws[gid].i[ 7];
|
|
|
|
const u32 pw_len = pws[gid].pw_len & 63;
|
|
|
|
/**
|
|
* loop
|
|
*/
|
|
|
|
for (u32 il_pos = 0; il_pos < IL_CNT; il_pos += VECT_SIZE)
|
|
{
|
|
u32x w0[4] = { 0 };
|
|
u32x w1[4] = { 0 };
|
|
u32x w2[4] = { 0 };
|
|
u32x w3[4] = { 0 };
|
|
|
|
const u32x out_len = apply_rules_vect_optimized (pw_buf0, pw_buf1, pw_len, rules_buf, il_pos, w0, w1);
|
|
|
|
append_0x01_2x4_VV (w0, w1, out_len);
|
|
|
|
/**
|
|
* GOST
|
|
*/
|
|
|
|
u32x w[16];
|
|
|
|
w[ 0] = w0[0];
|
|
w[ 1] = w0[1];
|
|
w[ 2] = w0[2];
|
|
w[ 3] = w0[3];
|
|
w[ 4] = w1[0];
|
|
w[ 5] = w1[1];
|
|
w[ 6] = w1[2];
|
|
w[ 7] = w1[3];
|
|
w[ 8] = 0;
|
|
w[ 9] = 0;
|
|
w[10] = 0;
|
|
w[11] = 0;
|
|
w[12] = 0;
|
|
w[13] = 0;
|
|
w[14] = 0;
|
|
w[15] = 0;
|
|
|
|
/**
|
|
* reverse message block
|
|
*/
|
|
|
|
u64x m[8];
|
|
|
|
m[0] = hl32_to_64 (w[15], w[14]);
|
|
m[1] = hl32_to_64 (w[13], w[12]);
|
|
m[2] = hl32_to_64 (w[11], w[10]);
|
|
m[3] = hl32_to_64 (w[ 9], w[ 8]);
|
|
m[4] = hl32_to_64 (w[ 7], w[ 6]);
|
|
m[5] = hl32_to_64 (w[ 5], w[ 4]);
|
|
m[6] = hl32_to_64 (w[ 3], w[ 2]);
|
|
m[7] = hl32_to_64 (w[ 1], w[ 0]);
|
|
|
|
m[0] = hc_swap64 (m[0]);
|
|
m[1] = hc_swap64 (m[1]);
|
|
m[2] = hc_swap64 (m[2]);
|
|
m[3] = hc_swap64 (m[3]);
|
|
m[4] = hc_swap64 (m[4]);
|
|
m[5] = hc_swap64 (m[5]);
|
|
m[6] = hc_swap64 (m[6]);
|
|
m[7] = hc_swap64 (m[7]);
|
|
|
|
// state buffer (hash)
|
|
|
|
u64x h[8];
|
|
|
|
h[0] = INITVAL;
|
|
h[1] = INITVAL;
|
|
h[2] = INITVAL;
|
|
h[3] = INITVAL;
|
|
h[4] = INITVAL;
|
|
h[5] = INITVAL;
|
|
h[6] = INITVAL;
|
|
h[7] = INITVAL;
|
|
|
|
streebog_g (h, m, s_sbob_sl64);
|
|
|
|
u64x z[8];
|
|
|
|
z[0] = 0;
|
|
z[1] = 0;
|
|
z[2] = 0;
|
|
z[3] = 0;
|
|
z[4] = 0;
|
|
z[5] = 0;
|
|
z[6] = 0;
|
|
z[7] = hc_swap64 ((u64) (pw_len * 8));
|
|
|
|
streebog_g (h, z, s_sbob_sl64);
|
|
streebog_g (h, m, s_sbob_sl64);
|
|
|
|
const u32x r0 = l32_from_64 (h[0]);
|
|
const u32x r1 = h32_from_64 (h[0]);
|
|
const u32x r2 = l32_from_64 (h[1]);
|
|
const u32x r3 = h32_from_64 (h[1]);
|
|
|
|
COMPARE_M_SIMD (r0, r1, r2, r3);
|
|
}
|
|
}
|
|
|
|
KERNEL_FQ KERNEL_FA void m11700_m08 (KERN_ATTR_RULES ())
|
|
{
|
|
}
|
|
|
|
KERNEL_FQ KERNEL_FA void m11700_m16 (KERN_ATTR_RULES ())
|
|
{
|
|
}
|
|
|
|
KERNEL_FQ KERNEL_FA void m11700_s04 (KERN_ATTR_RULES ())
|
|
{
|
|
/**
|
|
* modifier
|
|
*/
|
|
|
|
const u64 gid = get_global_id (0);
|
|
const u64 lid = get_local_id (0);
|
|
const u64 lsz = get_local_size (0);
|
|
|
|
/**
|
|
* shared lookup table
|
|
*/
|
|
|
|
LOCAL_VK u64 s_sbob_sl64[8][256];
|
|
|
|
for (u32 i = lid; i < 256; i += lsz)
|
|
{
|
|
s_sbob_sl64[0][i] = sbob256_sl64[0][i];
|
|
s_sbob_sl64[1][i] = sbob256_sl64[1][i];
|
|
s_sbob_sl64[2][i] = sbob256_sl64[2][i];
|
|
s_sbob_sl64[3][i] = sbob256_sl64[3][i];
|
|
s_sbob_sl64[4][i] = sbob256_sl64[4][i];
|
|
s_sbob_sl64[5][i] = sbob256_sl64[5][i];
|
|
s_sbob_sl64[6][i] = sbob256_sl64[6][i];
|
|
s_sbob_sl64[7][i] = sbob256_sl64[7][i];
|
|
}
|
|
|
|
SYNC_THREADS ();
|
|
|
|
if (gid >= GID_CNT) return;
|
|
|
|
/**
|
|
* base
|
|
*/
|
|
|
|
u32 pw_buf0[4];
|
|
u32 pw_buf1[4];
|
|
|
|
pw_buf0[0] = pws[gid].i[ 0];
|
|
pw_buf0[1] = pws[gid].i[ 1];
|
|
pw_buf0[2] = pws[gid].i[ 2];
|
|
pw_buf0[3] = pws[gid].i[ 3];
|
|
pw_buf1[0] = pws[gid].i[ 4];
|
|
pw_buf1[1] = pws[gid].i[ 5];
|
|
pw_buf1[2] = pws[gid].i[ 6];
|
|
pw_buf1[3] = pws[gid].i[ 7];
|
|
|
|
const u32 pw_len = pws[gid].pw_len & 63;
|
|
|
|
/**
|
|
* digest
|
|
*/
|
|
|
|
const u32 search[4] =
|
|
{
|
|
digests_buf[DIGESTS_OFFSET_HOST].digest_buf[DGST_R0],
|
|
digests_buf[DIGESTS_OFFSET_HOST].digest_buf[DGST_R1],
|
|
digests_buf[DIGESTS_OFFSET_HOST].digest_buf[DGST_R2],
|
|
digests_buf[DIGESTS_OFFSET_HOST].digest_buf[DGST_R3]
|
|
};
|
|
|
|
/**
|
|
* loop
|
|
*/
|
|
|
|
for (u32 il_pos = 0; il_pos < IL_CNT; il_pos += VECT_SIZE)
|
|
{
|
|
u32x w0[4] = { 0 };
|
|
u32x w1[4] = { 0 };
|
|
u32x w2[4] = { 0 };
|
|
u32x w3[4] = { 0 };
|
|
|
|
const u32x out_len = apply_rules_vect_optimized (pw_buf0, pw_buf1, pw_len, rules_buf, il_pos, w0, w1);
|
|
|
|
append_0x01_2x4_VV (w0, w1, out_len);
|
|
|
|
/**
|
|
* GOST
|
|
*/
|
|
|
|
u32x w[16];
|
|
|
|
w[ 0] = w0[0];
|
|
w[ 1] = w0[1];
|
|
w[ 2] = w0[2];
|
|
w[ 3] = w0[3];
|
|
w[ 4] = w1[0];
|
|
w[ 5] = w1[1];
|
|
w[ 6] = w1[2];
|
|
w[ 7] = w1[3];
|
|
w[ 8] = 0;
|
|
w[ 9] = 0;
|
|
w[10] = 0;
|
|
w[11] = 0;
|
|
w[12] = 0;
|
|
w[13] = 0;
|
|
w[14] = 0;
|
|
w[15] = 0;
|
|
|
|
/**
|
|
* reverse message block
|
|
*/
|
|
|
|
u64x m[8];
|
|
|
|
m[0] = hl32_to_64 (w[15], w[14]);
|
|
m[1] = hl32_to_64 (w[13], w[12]);
|
|
m[2] = hl32_to_64 (w[11], w[10]);
|
|
m[3] = hl32_to_64 (w[ 9], w[ 8]);
|
|
m[4] = hl32_to_64 (w[ 7], w[ 6]);
|
|
m[5] = hl32_to_64 (w[ 5], w[ 4]);
|
|
m[6] = hl32_to_64 (w[ 3], w[ 2]);
|
|
m[7] = hl32_to_64 (w[ 1], w[ 0]);
|
|
|
|
m[0] = hc_swap64 (m[0]);
|
|
m[1] = hc_swap64 (m[1]);
|
|
m[2] = hc_swap64 (m[2]);
|
|
m[3] = hc_swap64 (m[3]);
|
|
m[4] = hc_swap64 (m[4]);
|
|
m[5] = hc_swap64 (m[5]);
|
|
m[6] = hc_swap64 (m[6]);
|
|
m[7] = hc_swap64 (m[7]);
|
|
|
|
// state buffer (hash)
|
|
|
|
u64x h[8];
|
|
|
|
h[0] = INITVAL;
|
|
h[1] = INITVAL;
|
|
h[2] = INITVAL;
|
|
h[3] = INITVAL;
|
|
h[4] = INITVAL;
|
|
h[5] = INITVAL;
|
|
h[6] = INITVAL;
|
|
h[7] = INITVAL;
|
|
|
|
streebog_g (h, m, s_sbob_sl64);
|
|
|
|
u64x z[8];
|
|
|
|
z[0] = 0;
|
|
z[1] = 0;
|
|
z[2] = 0;
|
|
z[3] = 0;
|
|
z[4] = 0;
|
|
z[5] = 0;
|
|
z[6] = 0;
|
|
z[7] = hc_swap64 ((u64) (pw_len * 8));
|
|
|
|
streebog_g (h, z, s_sbob_sl64);
|
|
streebog_g (h, m, s_sbob_sl64);
|
|
|
|
const u32x r0 = l32_from_64 (h[0]);
|
|
const u32x r1 = h32_from_64 (h[0]);
|
|
const u32x r2 = l32_from_64 (h[1]);
|
|
const u32x r3 = h32_from_64 (h[1]);
|
|
|
|
COMPARE_S_SIMD (r0, r1, r2, r3);
|
|
}
|
|
}
|
|
|
|
KERNEL_FQ KERNEL_FA void m11700_s08 (KERN_ATTR_RULES ())
|
|
{
|
|
}
|
|
|
|
KERNEL_FQ KERNEL_FA void m11700_s16 (KERN_ATTR_RULES ())
|
|
{
|
|
}
|