X-Git-Url: https://www.flypig.org.uk/git/?a=blobdiff_plain;f=OpenCL%2Fm06800.cl;h=6152ed9a1da0aa8bdb4960c8bd8c1caf688ef74f;hb=161a6eb4bc643d8e636e96eda613f5137d30da59;hp=2aadf6f5becc4f25374b2dde5b78d7f0304d9b3c;hpb=331188167cc7b4f6f2a336d63daca23fbdfc199c;p=hashcat.git diff --git a/OpenCL/m06800.cl b/OpenCL/m06800.cl index 2aadf6f..6152ed9 100644 --- a/OpenCL/m06800.cl +++ b/OpenCL/m06800.cl @@ -1,24 +1,20 @@ /** - * Author......: Jens Steube + * Authors.....: Jens Steube + * Gabriele Gristina + * * License.....: MIT */ #define _SHA256_ -#include "include/constants.h" -#include "include/kernel_vendor.h" +#include "inc_vendor.cl" +#include "inc_hash_constants.h" +#include "inc_hash_functions.cl" +#include "inc_types.cl" +#include "inc_common.cl" -#define DGST_R0 0 -#define DGST_R1 1 -#define DGST_R2 2 -#define DGST_R3 3 - -#include "include/kernel_functions.c" -#include "OpenCL/types_ocl.c" -#include "OpenCL/common.c" - -#define COMPARE_S "OpenCL/check_single_comp4.c" -#define COMPARE_M "OpenCL/check_multi_comp4.c" +#define COMPARE_S "inc_comp_single.cl" +#define COMPARE_M "inc_comp_multi.cl" __constant u32 te0[256] = { @@ -707,7 +703,7 @@ __constant u32 rcon[] = 0x1b000000, 0x36000000, }; -static void AES256_ExpandKey (u32 *userkey, u32 *rek, __local u32 s_te0[256], __local u32 s_te1[256], __local u32 s_te2[256], __local u32 s_te3[256], __local u32 s_te4[256]) +void AES256_ExpandKey (u32 *userkey, u32 *rek, __local u32 *s_te0, __local u32 *s_te1, __local u32 *s_te2, __local u32 *s_te3, __local u32 *s_te4) { rek[0] = userkey[0]; rek[1] = userkey[1]; @@ -763,7 +759,7 @@ static void AES256_ExpandKey (u32 *userkey, u32 *rek, __local u32 s_te0[256], __ } } -static void AES256_InvertKey (u32 *rdk, __local u32 s_td0[256], __local u32 s_td1[256], __local u32 s_td2[256], __local u32 s_td3[256], __local u32 s_td4[256], __local u32 s_te0[256], __local u32 s_te1[256], __local u32 s_te2[256], __local u32 s_te3[256], __local u32 s_te4[256]) +void AES256_InvertKey (u32 *rdk, __local u32 *s_td0, __local u32 *s_td1, __local u32 *s_td2, __local u32 *s_td3, __local u32 *s_td4, __local u32 *s_te0, __local u32 *s_te1, __local u32 *s_te2, __local u32 *s_te3, __local u32 *s_te4) { for (u32 i = 0, j = 56; i < j; i += 4, j -= 4) { @@ -803,7 +799,7 @@ static void AES256_InvertKey (u32 *rdk, __local u32 s_td0[256], __local u32 s_td } } -static void AES256_decrypt (const u32 *in, u32 *out, const u32 *rdk, __local u32 s_td0[256], __local u32 s_td1[256], __local u32 s_td2[256], __local u32 s_td3[256], __local u32 s_td4[256]) +void AES256_decrypt (const u32 *in, u32 *out, const u32 *rdk, __local u32 *s_td0, __local u32 *s_td1, __local u32 *s_td2, __local u32 *s_td3, __local u32 *s_td4) { u32 s0 = in[0] ^ rdk[0]; u32 s1 = in[1] ^ rdk[1]; @@ -893,7 +889,7 @@ static void AES256_decrypt (const u32 *in, u32 *out, const u32 *rdk, __local u32 ^ rdk[59]; } -static void AES256_encrypt (const u32 *in, u32 *out, const u32 *rek, __local u32 s_te0[256], __local u32 s_te1[256], __local u32 s_te2[256], __local u32 s_te3[256], __local u32 s_te4[256]) +void AES256_encrypt (const u32 *in, u32 *out, const u32 *rek, __local u32 *s_te0, __local u32 *s_te1, __local u32 *s_te2, __local u32 *s_te3, __local u32 *s_te4) { u32 s0 = in[0] ^ rek[0]; u32 s1 = in[1] ^ rek[1]; @@ -1003,7 +999,7 @@ __constant u32 k_sha256[64] = SHA256C3c, SHA256C3d, SHA256C3e, SHA256C3f, }; -static void sha256_transform (const u32 w0[4], const u32 w1[4], const u32 w2[4], const u32 w3[4], u32 digest[8]) +void sha256_transform (const u32 w0[4], const u32 w1[4], const u32 w2[4], const u32 w3[4], u32 digest[8]) { u32 a = digest[0]; u32 b = digest[1]; @@ -1073,7 +1069,9 @@ static void sha256_transform (const u32 w0[4], const u32 w1[4], const u32 w2[4], ROUND_STEP (0); + #ifdef _unroll #pragma unroll + #endif for (int i = 16; i < 64; i += 16) { ROUND_EXPAND (); ROUND_STEP (i); @@ -1089,7 +1087,7 @@ static void sha256_transform (const u32 w0[4], const u32 w1[4], const u32 w2[4], digest[7] += h; } -static void hmac_sha256_pad (u32 w0[4], u32 w1[4], u32 w2[4], u32 w3[4], u32 ipad[8], u32 opad[8]) +void hmac_sha256_pad (u32 w0[4], u32 w1[4], u32 w2[4], u32 w3[4], u32 ipad[8], u32 opad[8]) { w0[0] = w0[0] ^ 0x36363636; w0[1] = w0[1] ^ 0x36363636; @@ -1148,7 +1146,7 @@ static void hmac_sha256_pad (u32 w0[4], u32 w1[4], u32 w2[4], u32 w3[4], u32 ipa sha256_transform (w0, w1, w2, w3, opad); } -static void hmac_sha256_run (u32 w0[4], u32 w1[4], u32 w2[4], u32 w3[4], u32 ipad[8], u32 opad[8], u32 digest[8]) +void hmac_sha256_run (u32 w0[4], u32 w1[4], u32 w2[4], u32 w3[4], u32 ipad[8], u32 opad[8], u32 digest[8]) { digest[0] = ipad[0]; digest[1] = ipad[1]; @@ -1190,7 +1188,7 @@ static void hmac_sha256_run (u32 w0[4], u32 w1[4], u32 w2[4], u32 w3[4], u32 ipa sha256_transform (w0, w1, w2, w3, digest); } -__kernel void __attribute__((reqd_work_group_size (64, 1, 1))) m06800_init (__global pw_t *pws, __global kernel_rule_t *rules_buf, __global comb_t *combs_buf, __global bf_t *bfs_buf, __global lastpass_tmp_t *tmps, __global void *hooks, __global u32 *bitmaps_buf_s1_a, __global u32 *bitmaps_buf_s1_b, __global u32 *bitmaps_buf_s1_c, __global u32 *bitmaps_buf_s1_d, __global u32 *bitmaps_buf_s2_a, __global u32 *bitmaps_buf_s2_b, __global u32 *bitmaps_buf_s2_c, __global u32 *bitmaps_buf_s2_d, __global plain_t *plains_buf, __global digest_t *digests_buf, __global u32 *hashes_shown, __global salt_t *salt_bufs, __global void *esalt_bufs, __global u32 *d_return_buf, __global u32 *d_scryptV_buf, const u32 bitmap_mask, const u32 bitmap_shift1, const u32 bitmap_shift2, const u32 salt_pos, const u32 loop_pos, const u32 loop_cnt, const u32 rules_cnt, const u32 digests_cnt, const u32 digests_offset, const u32 combs_mode, const u32 gid_max) +__kernel void m06800_init (__global pw_t *pws, __global kernel_rule_t *rules_buf, __global comb_t *combs_buf, __global bf_t *bfs_buf, __global lastpass_tmp_t *tmps, __global void *hooks, __global u32 *bitmaps_buf_s1_a, __global u32 *bitmaps_buf_s1_b, __global u32 *bitmaps_buf_s1_c, __global u32 *bitmaps_buf_s1_d, __global u32 *bitmaps_buf_s2_a, __global u32 *bitmaps_buf_s2_b, __global u32 *bitmaps_buf_s2_c, __global u32 *bitmaps_buf_s2_d, __global plain_t *plains_buf, __global digest_t *digests_buf, __global u32 *hashes_shown, __global salt_t *salt_bufs, __global void *esalt_bufs, __global u32 *d_return_buf, __global u32 *d_scryptV0_buf, __global u32 *d_scryptV1_buf, __global u32 *d_scryptV2_buf, __global u32 *d_scryptV3_buf, const u32 bitmap_mask, const u32 bitmap_shift1, const u32 bitmap_shift2, const u32 salt_pos, const u32 loop_pos, const u32 loop_cnt, const u32 il_cnt, const u32 digests_cnt, const u32 digests_offset, const u32 combs_mode, const u32 gid_max) { /** * base @@ -1350,7 +1348,7 @@ __kernel void __attribute__((reqd_work_group_size (64, 1, 1))) m06800_init (__gl tmps[gid].out[7] = dgst[7]; } -__kernel void __attribute__((reqd_work_group_size (64, 1, 1))) m06800_loop (__global pw_t *pws, __global kernel_rule_t *rules_buf, __global comb_t *combs_buf, __global bf_t *bfs_buf, __global lastpass_tmp_t *tmps, __global void *hooks, __global u32 *bitmaps_buf_s1_a, __global u32 *bitmaps_buf_s1_b, __global u32 *bitmaps_buf_s1_c, __global u32 *bitmaps_buf_s1_d, __global u32 *bitmaps_buf_s2_a, __global u32 *bitmaps_buf_s2_b, __global u32 *bitmaps_buf_s2_c, __global u32 *bitmaps_buf_s2_d, __global plain_t *plains_buf, __global digest_t *digests_buf, __global u32 *hashes_shown, __global salt_t *salt_bufs, __global void *esalt_bufs, __global u32 *d_return_buf, __global u32 *d_scryptV_buf, const u32 bitmap_mask, const u32 bitmap_shift1, const u32 bitmap_shift2, const u32 salt_pos, const u32 loop_pos, const u32 loop_cnt, const u32 rules_cnt, const u32 digests_cnt, const u32 digests_offset, const u32 combs_mode, const u32 gid_max) +__kernel void m06800_loop (__global pw_t *pws, __global kernel_rule_t *rules_buf, __global comb_t *combs_buf, __global bf_t *bfs_buf, __global lastpass_tmp_t *tmps, __global void *hooks, __global u32 *bitmaps_buf_s1_a, __global u32 *bitmaps_buf_s1_b, __global u32 *bitmaps_buf_s1_c, __global u32 *bitmaps_buf_s1_d, __global u32 *bitmaps_buf_s2_a, __global u32 *bitmaps_buf_s2_b, __global u32 *bitmaps_buf_s2_c, __global u32 *bitmaps_buf_s2_d, __global plain_t *plains_buf, __global digest_t *digests_buf, __global u32 *hashes_shown, __global salt_t *salt_bufs, __global void *esalt_bufs, __global u32 *d_return_buf, __global u32 *d_scryptV0_buf, __global u32 *d_scryptV1_buf, __global u32 *d_scryptV2_buf, __global u32 *d_scryptV3_buf, const u32 bitmap_mask, const u32 bitmap_shift1, const u32 bitmap_shift2, const u32 salt_pos, const u32 loop_pos, const u32 loop_cnt, const u32 il_cnt, const u32 digests_cnt, const u32 digests_offset, const u32 combs_mode, const u32 gid_max) { const u32 gid = get_global_id (0); @@ -1453,7 +1451,7 @@ __kernel void __attribute__((reqd_work_group_size (64, 1, 1))) m06800_loop (__gl tmps[gid].out[7] = out[7]; } -__kernel void __attribute__((reqd_work_group_size (64, 1, 1))) m06800_comp (__global pw_t *pws, __global kernel_rule_t *rules_buf, __global comb_t *combs_buf, __global bf_t *bfs_buf, __global lastpass_tmp_t *tmps, __global void *hooks, __global u32 *bitmaps_buf_s1_a, __global u32 *bitmaps_buf_s1_b, __global u32 *bitmaps_buf_s1_c, __global u32 *bitmaps_buf_s1_d, __global u32 *bitmaps_buf_s2_a, __global u32 *bitmaps_buf_s2_b, __global u32 *bitmaps_buf_s2_c, __global u32 *bitmaps_buf_s2_d, __global plain_t *plains_buf, __global digest_t *digests_buf, __global u32 *hashes_shown, __global salt_t *salt_bufs, __global void *esalt_bufs, __global u32 *d_return_buf, __global u32 *d_scryptV_buf, const u32 bitmap_mask, const u32 bitmap_shift1, const u32 bitmap_shift2, const u32 salt_pos, const u32 loop_pos, const u32 loop_cnt, const u32 rules_cnt, const u32 digests_cnt, const u32 digests_offset, const u32 combs_mode, const u32 gid_max) +__kernel void m06800_comp (__global pw_t *pws, __global kernel_rule_t *rules_buf, __global comb_t *combs_buf, __global bf_t *bfs_buf, __global lastpass_tmp_t *tmps, __global void *hooks, __global u32 *bitmaps_buf_s1_a, __global u32 *bitmaps_buf_s1_b, __global u32 *bitmaps_buf_s1_c, __global u32 *bitmaps_buf_s1_d, __global u32 *bitmaps_buf_s2_a, __global u32 *bitmaps_buf_s2_b, __global u32 *bitmaps_buf_s2_c, __global u32 *bitmaps_buf_s2_d, __global plain_t *plains_buf, __global digest_t *digests_buf, __global u32 *hashes_shown, __global salt_t *salt_bufs, __global void *esalt_bufs, __global u32 *d_return_buf, __global u32 *d_scryptV0_buf, __global u32 *d_scryptV1_buf, __global u32 *d_scryptV2_buf, __global u32 *d_scryptV3_buf, const u32 bitmap_mask, const u32 bitmap_shift1, const u32 bitmap_shift2, const u32 salt_pos, const u32 loop_pos, const u32 loop_cnt, const u32 il_cnt, const u32 digests_cnt, const u32 digests_offset, const u32 combs_mode, const u32 gid_max) { /** * base @@ -1461,25 +1459,12 @@ __kernel void __attribute__((reqd_work_group_size (64, 1, 1))) m06800_comp (__gl const u32 gid = get_global_id (0); const u32 lid = get_local_id (0); - - /** - * salt - */ - - const u32 data[4] = - { - digests_buf[digests_offset].digest_buf[0], - digests_buf[digests_offset].digest_buf[1], - digests_buf[digests_offset].digest_buf[2], - digests_buf[digests_offset].digest_buf[3], - }; + const u32 lsz = get_local_size (0); /** * aes shared */ - const u32 lid4 = lid * 4; - __local u32 s_td0[256]; __local u32 s_td1[256]; __local u32 s_td2[256]; @@ -1492,60 +1477,37 @@ __kernel void __attribute__((reqd_work_group_size (64, 1, 1))) m06800_comp (__gl __local u32 s_te3[256]; __local u32 s_te4[256]; - s_td0[lid4 + 0] = td0[lid4 + 0]; - s_td0[lid4 + 1] = td0[lid4 + 1]; - s_td0[lid4 + 2] = td0[lid4 + 2]; - s_td0[lid4 + 3] = td0[lid4 + 3]; - - s_td1[lid4 + 0] = td1[lid4 + 0]; - s_td1[lid4 + 1] = td1[lid4 + 1]; - s_td1[lid4 + 2] = td1[lid4 + 2]; - s_td1[lid4 + 3] = td1[lid4 + 3]; - - s_td2[lid4 + 0] = td2[lid4 + 0]; - s_td2[lid4 + 1] = td2[lid4 + 1]; - s_td2[lid4 + 2] = td2[lid4 + 2]; - s_td2[lid4 + 3] = td2[lid4 + 3]; - - s_td3[lid4 + 0] = td3[lid4 + 0]; - s_td3[lid4 + 1] = td3[lid4 + 1]; - s_td3[lid4 + 2] = td3[lid4 + 2]; - s_td3[lid4 + 3] = td3[lid4 + 3]; - - s_td4[lid4 + 0] = td4[lid4 + 0]; - s_td4[lid4 + 1] = td4[lid4 + 1]; - s_td4[lid4 + 2] = td4[lid4 + 2]; - s_td4[lid4 + 3] = td4[lid4 + 3]; - - s_te0[lid4 + 0] = te0[lid4 + 0]; - s_te0[lid4 + 1] = te0[lid4 + 1]; - s_te0[lid4 + 2] = te0[lid4 + 2]; - s_te0[lid4 + 3] = te0[lid4 + 3]; - - s_te1[lid4 + 0] = te1[lid4 + 0]; - s_te1[lid4 + 1] = te1[lid4 + 1]; - s_te1[lid4 + 2] = te1[lid4 + 2]; - s_te1[lid4 + 3] = te1[lid4 + 3]; - - s_te2[lid4 + 0] = te2[lid4 + 0]; - s_te2[lid4 + 1] = te2[lid4 + 1]; - s_te2[lid4 + 2] = te2[lid4 + 2]; - s_te2[lid4 + 3] = te2[lid4 + 3]; - - s_te3[lid4 + 0] = te3[lid4 + 0]; - s_te3[lid4 + 1] = te3[lid4 + 1]; - s_te3[lid4 + 2] = te3[lid4 + 2]; - s_te3[lid4 + 3] = te3[lid4 + 3]; - - s_te4[lid4 + 0] = te4[lid4 + 0]; - s_te4[lid4 + 1] = te4[lid4 + 1]; - s_te4[lid4 + 2] = te4[lid4 + 2]; - s_te4[lid4 + 3] = te4[lid4 + 3]; + for (u32 i = lid; i < 256; i += lsz) + { + s_td0[i] = td0[i]; + s_td1[i] = td1[i]; + s_td2[i] = td2[i]; + s_td3[i] = td3[i]; + s_td4[i] = td4[i]; + + s_te0[i] = te0[i]; + s_te1[i] = te1[i]; + s_te2[i] = te2[i]; + s_te3[i] = te3[i]; + s_te4[i] = te4[i]; + } barrier (CLK_LOCAL_MEM_FENCE); if (gid >= gid_max) return; + /** + * salt + */ + + const u32 data[4] = + { + digests_buf[digests_offset].digest_buf[0], + digests_buf[digests_offset].digest_buf[1], + digests_buf[digests_offset].digest_buf[2], + digests_buf[digests_offset].digest_buf[3], + }; + /** * aes init */ @@ -1575,7 +1537,9 @@ __kernel void __attribute__((reqd_work_group_size (64, 1, 1))) m06800_comp (__gl u32 rdk[KEYLEN]; - #pragma unroll 60 + #ifdef _unroll + #pragma unroll + #endif for (u32 i = 0; i < KEYLEN; i++) rdk[i] = rek[i]; AES256_InvertKey (rdk, s_td0, s_td1, s_td2, s_td3, s_td4, s_te0, s_te1, s_te2, s_te3, s_te4); @@ -1603,9 +1567,7 @@ __kernel void __attribute__((reqd_work_group_size (64, 1, 1))) m06800_comp (__gl && (out[2] == salt_buf[2]) && (out[3] == salt_buf[3])) { - mark_hash (plains_buf, hashes_shown, digests_offset + 0, gid, 0); - - d_return_buf[lid] = 1; + mark_hash (plains_buf, d_return_buf, salt_pos, 0, digests_offset + 0, gid, 0); } /**