/** * Author......: See docs/credits.txt * License.....: MIT */ #define NEW_SIMD_CODE #define XSTR(x) #x #define STR(x) XSTR(x) #ifdef KERNEL_STATIC #include STR(INCLUDE_PATH/inc_vendor.h) #include STR(INCLUDE_PATH/inc_types.h) #include STR(INCLUDE_PATH/inc_platform.cl) #include STR(INCLUDE_PATH/inc_common.cl) #include STR(INCLUDE_PATH/inc_rp_optimized.h) #include STR(INCLUDE_PATH/inc_rp_optimized.cl) #include STR(INCLUDE_PATH/inc_simd.cl) #include STR(INCLUDE_PATH/inc_hash_whirlpool.cl) #endif DECLSPEC void whirlpool_transform_transport_vector (const u32x *w, u32x *digest, SHM_TYPE u64 *s_MT0, SHM_TYPE u64 *s_MT1, SHM_TYPE u64 *s_MT2, SHM_TYPE u64 *s_MT3, SHM_TYPE u64 *s_MT4, SHM_TYPE u64 *s_MT5, SHM_TYPE u64 *s_MT6, SHM_TYPE u64 *s_MT7) { whirlpool_transform_vector (w + 0, w + 4, w + 8, w + 12, digest, s_MT0, s_MT1, s_MT2, s_MT3, s_MT4, s_MT5, s_MT6, s_MT7); } KERNEL_FQ void m06100_m04 (KERN_ATTR_RULES ()) { /** * modifier */ const u64 gid = get_global_id (0); const u64 lid = get_local_id (0); const u64 lsz = get_local_size (0); /** * Whirlpool shared */ #ifdef REAL_SHM LOCAL_VK u64 s_MT0[256]; LOCAL_VK u64 s_MT1[256]; LOCAL_VK u64 s_MT2[256]; LOCAL_VK u64 s_MT3[256]; LOCAL_VK u64 s_MT4[256]; LOCAL_VK u64 s_MT5[256]; LOCAL_VK u64 s_MT6[256]; LOCAL_VK u64 s_MT7[256]; for (u32 i = lid; i < 256; i += lsz) { s_MT0[i] = MT0[i]; s_MT1[i] = MT1[i]; s_MT2[i] = MT2[i]; s_MT3[i] = MT3[i]; s_MT4[i] = MT4[i]; s_MT5[i] = MT5[i]; s_MT6[i] = MT6[i]; s_MT7[i] = MT7[i]; } SYNC_THREADS (); #else CONSTANT_AS u64a *s_MT0 = MT0; CONSTANT_AS u64a *s_MT1 = MT1; CONSTANT_AS u64a *s_MT2 = MT2; CONSTANT_AS u64a *s_MT3 = MT3; CONSTANT_AS u64a *s_MT4 = MT4; CONSTANT_AS u64a *s_MT5 = MT5; CONSTANT_AS u64a *s_MT6 = MT6; CONSTANT_AS u64a *s_MT7 = MT7; #endif if (gid >= GID_CNT) return; /** * base */ u32 pw_buf0[4]; u32 pw_buf1[4]; pw_buf0[0] = pws[gid].i[0]; pw_buf0[1] = pws[gid].i[1]; pw_buf0[2] = pws[gid].i[2]; pw_buf0[3] = pws[gid].i[3]; pw_buf1[0] = pws[gid].i[4]; pw_buf1[1] = pws[gid].i[5]; pw_buf1[2] = pws[gid].i[6]; pw_buf1[3] = pws[gid].i[7]; const u32 pw_len = pws[gid].pw_len & 63; /** * loop */ for (u32 il_pos = 0; il_pos < IL_CNT; il_pos += VECT_SIZE) { u32x w0[4] = { 0 }; u32x w1[4] = { 0 }; u32x w2[4] = { 0 }; u32x w3[4] = { 0 }; const u32x out_len = apply_rules_vect_optimized (pw_buf0, pw_buf1, pw_len, rules_buf, il_pos, w0, w1); append_0x80_2x4_VV (w0, w1, out_len); u32x w[16]; w[ 0] = hc_swap32 (w0[0]); w[ 1] = hc_swap32 (w0[1]); w[ 2] = hc_swap32 (w0[2]); w[ 3] = hc_swap32 (w0[3]); w[ 4] = hc_swap32 (w1[0]); w[ 5] = hc_swap32 (w1[1]); w[ 6] = hc_swap32 (w1[2]); w[ 7] = hc_swap32 (w1[3]); w[ 8] = 0; w[ 9] = 0; w[10] = 0; w[11] = 0; w[12] = 0; w[13] = 0; w[14] = 0; w[15] = out_len * 8; /** * Whirlpool */ u32x dgst[16]; dgst[ 0] = 0; dgst[ 1] = 0; dgst[ 2] = 0; dgst[ 3] = 0; dgst[ 4] = 0; dgst[ 5] = 0; dgst[ 6] = 0; dgst[ 7] = 0; dgst[ 8] = 0; dgst[ 9] = 0; dgst[10] = 0; dgst[11] = 0; dgst[12] = 0; dgst[13] = 0; dgst[14] = 0; dgst[15] = 0; whirlpool_transform_transport_vector (w, dgst, s_MT0, s_MT1, s_MT2, s_MT3, s_MT4, s_MT5, s_MT6, s_MT7); COMPARE_M_SIMD (dgst[0], dgst[1], dgst[2], dgst[3]); } } KERNEL_FQ void m06100_m08 (KERN_ATTR_RULES ()) { } KERNEL_FQ void m06100_m16 (KERN_ATTR_RULES ()) { } KERNEL_FQ void m06100_s04 (KERN_ATTR_RULES ()) { /** * modifier */ const u64 gid = get_global_id (0); const u64 lid = get_local_id (0); const u64 lsz = get_local_size (0); /** * Whirlpool shared */ #ifdef REAL_SHM LOCAL_VK u64 s_MT0[256]; LOCAL_VK u64 s_MT1[256]; LOCAL_VK u64 s_MT2[256]; LOCAL_VK u64 s_MT3[256]; LOCAL_VK u64 s_MT4[256]; LOCAL_VK u64 s_MT5[256]; LOCAL_VK u64 s_MT6[256]; LOCAL_VK u64 s_MT7[256]; for (u32 i = lid; i < 256; i += lsz) { s_MT0[i] = MT0[i]; s_MT1[i] = MT1[i]; s_MT2[i] = MT2[i]; s_MT3[i] = MT3[i]; s_MT4[i] = MT4[i]; s_MT5[i] = MT5[i]; s_MT6[i] = MT6[i]; s_MT7[i] = MT7[i]; } SYNC_THREADS (); #else CONSTANT_AS u64a *s_MT0 = MT0; CONSTANT_AS u64a *s_MT1 = MT1; CONSTANT_AS u64a *s_MT2 = MT2; CONSTANT_AS u64a *s_MT3 = MT3; CONSTANT_AS u64a *s_MT4 = MT4; CONSTANT_AS u64a *s_MT5 = MT5; CONSTANT_AS u64a *s_MT6 = MT6; CONSTANT_AS u64a *s_MT7 = MT7; #endif if (gid >= GID_CNT) return; /** * base */ u32 pw_buf0[4]; u32 pw_buf1[4]; pw_buf0[0] = pws[gid].i[0]; pw_buf0[1] = pws[gid].i[1]; pw_buf0[2] = pws[gid].i[2]; pw_buf0[3] = pws[gid].i[3]; pw_buf1[0] = pws[gid].i[4]; pw_buf1[1] = pws[gid].i[5]; pw_buf1[2] = pws[gid].i[6]; pw_buf1[3] = pws[gid].i[7]; const u32 pw_len = pws[gid].pw_len & 63; /** * digest */ const u32 search[4] = { digests_buf[DIGESTS_OFFSET_HOST].digest_buf[DGST_R0], digests_buf[DIGESTS_OFFSET_HOST].digest_buf[DGST_R1], digests_buf[DIGESTS_OFFSET_HOST].digest_buf[DGST_R2], digests_buf[DIGESTS_OFFSET_HOST].digest_buf[DGST_R3] }; /** * loop */ for (u32 il_pos = 0; il_pos < IL_CNT; il_pos += VECT_SIZE) { u32x w0[4] = { 0 }; u32x w1[4] = { 0 }; u32x w2[4] = { 0 }; u32x w3[4] = { 0 }; const u32x out_len = apply_rules_vect_optimized (pw_buf0, pw_buf1, pw_len, rules_buf, il_pos, w0, w1); append_0x80_2x4_VV (w0, w1, out_len); u32x w[16]; w[ 0] = hc_swap32 (w0[0]); w[ 1] = hc_swap32 (w0[1]); w[ 2] = hc_swap32 (w0[2]); w[ 3] = hc_swap32 (w0[3]); w[ 4] = hc_swap32 (w1[0]); w[ 5] = hc_swap32 (w1[1]); w[ 6] = hc_swap32 (w1[2]); w[ 7] = hc_swap32 (w1[3]); w[ 8] = 0; w[ 9] = 0; w[10] = 0; w[11] = 0; w[12] = 0; w[13] = 0; w[14] = 0; w[15] = out_len * 8; /** * Whirlpool */ u32x dgst[16]; dgst[ 0] = 0; dgst[ 1] = 0; dgst[ 2] = 0; dgst[ 3] = 0; dgst[ 4] = 0; dgst[ 5] = 0; dgst[ 6] = 0; dgst[ 7] = 0; dgst[ 8] = 0; dgst[ 9] = 0; dgst[10] = 0; dgst[11] = 0; dgst[12] = 0; dgst[13] = 0; dgst[14] = 0; dgst[15] = 0; whirlpool_transform_transport_vector (w, dgst, s_MT0, s_MT1, s_MT2, s_MT3, s_MT4, s_MT5, s_MT6, s_MT7); COMPARE_S_SIMD (dgst[0], dgst[1], dgst[2], dgst[3]); } } KERNEL_FQ void m06100_s08 (KERN_ATTR_RULES ()) { } KERNEL_FQ void m06100_s16 (KERN_ATTR_RULES ()) { }