2 * Author......: Jens Steube <jens.steube@gmail.com>
8 #include "include/constants.h"
9 #include "include/kernel_vendor.h"
28 #include "include/kernel_functions.c"
29 #include "types_amd.c"
30 #include "common_amd.c"
33 #define VECT_COMPARE_S "check_single_vect1_comp4.c"
34 #define VECT_COMPARE_M "check_multi_vect1_comp4.c"
38 #define VECT_COMPARE_S "check_single_vect4_comp4.c"
39 #define VECT_COMPARE_M "check_multi_vect4_comp4.c"
42 static void md5_transform (const u32x w0[4], const u32x w1[4], const u32x w2[4], const u32x w3[4], u32x digest[4])
66 MD5_STEP (MD5_Fo, a, b, c, d, w0_t, MD5C00, MD5S00);
67 MD5_STEP (MD5_Fo, d, a, b, c, w1_t, MD5C01, MD5S01);
68 MD5_STEP (MD5_Fo, c, d, a, b, w2_t, MD5C02, MD5S02);
69 MD5_STEP (MD5_Fo, b, c, d, a, w3_t, MD5C03, MD5S03);
70 MD5_STEP (MD5_Fo, a, b, c, d, w4_t, MD5C04, MD5S00);
71 MD5_STEP (MD5_Fo, d, a, b, c, w5_t, MD5C05, MD5S01);
72 MD5_STEP (MD5_Fo, c, d, a, b, w6_t, MD5C06, MD5S02);
73 MD5_STEP (MD5_Fo, b, c, d, a, w7_t, MD5C07, MD5S03);
74 MD5_STEP (MD5_Fo, a, b, c, d, w8_t, MD5C08, MD5S00);
75 MD5_STEP (MD5_Fo, d, a, b, c, w9_t, MD5C09, MD5S01);
76 MD5_STEP (MD5_Fo, c, d, a, b, wa_t, MD5C0a, MD5S02);
77 MD5_STEP (MD5_Fo, b, c, d, a, wb_t, MD5C0b, MD5S03);
78 MD5_STEP (MD5_Fo, a, b, c, d, wc_t, MD5C0c, MD5S00);
79 MD5_STEP (MD5_Fo, d, a, b, c, wd_t, MD5C0d, MD5S01);
80 MD5_STEP (MD5_Fo, c, d, a, b, we_t, MD5C0e, MD5S02);
81 MD5_STEP (MD5_Fo, b, c, d, a, wf_t, MD5C0f, MD5S03);
83 MD5_STEP (MD5_Go, a, b, c, d, w1_t, MD5C10, MD5S10);
84 MD5_STEP (MD5_Go, d, a, b, c, w6_t, MD5C11, MD5S11);
85 MD5_STEP (MD5_Go, c, d, a, b, wb_t, MD5C12, MD5S12);
86 MD5_STEP (MD5_Go, b, c, d, a, w0_t, MD5C13, MD5S13);
87 MD5_STEP (MD5_Go, a, b, c, d, w5_t, MD5C14, MD5S10);
88 MD5_STEP (MD5_Go, d, a, b, c, wa_t, MD5C15, MD5S11);
89 MD5_STEP (MD5_Go, c, d, a, b, wf_t, MD5C16, MD5S12);
90 MD5_STEP (MD5_Go, b, c, d, a, w4_t, MD5C17, MD5S13);
91 MD5_STEP (MD5_Go, a, b, c, d, w9_t, MD5C18, MD5S10);
92 MD5_STEP (MD5_Go, d, a, b, c, we_t, MD5C19, MD5S11);
93 MD5_STEP (MD5_Go, c, d, a, b, w3_t, MD5C1a, MD5S12);
94 MD5_STEP (MD5_Go, b, c, d, a, w8_t, MD5C1b, MD5S13);
95 MD5_STEP (MD5_Go, a, b, c, d, wd_t, MD5C1c, MD5S10);
96 MD5_STEP (MD5_Go, d, a, b, c, w2_t, MD5C1d, MD5S11);
97 MD5_STEP (MD5_Go, c, d, a, b, w7_t, MD5C1e, MD5S12);
98 MD5_STEP (MD5_Go, b, c, d, a, wc_t, MD5C1f, MD5S13);
100 MD5_STEP (MD5_H , a, b, c, d, w5_t, MD5C20, MD5S20);
101 MD5_STEP (MD5_H , d, a, b, c, w8_t, MD5C21, MD5S21);
102 MD5_STEP (MD5_H , c, d, a, b, wb_t, MD5C22, MD5S22);
103 MD5_STEP (MD5_H , b, c, d, a, we_t, MD5C23, MD5S23);
104 MD5_STEP (MD5_H , a, b, c, d, w1_t, MD5C24, MD5S20);
105 MD5_STEP (MD5_H , d, a, b, c, w4_t, MD5C25, MD5S21);
106 MD5_STEP (MD5_H , c, d, a, b, w7_t, MD5C26, MD5S22);
107 MD5_STEP (MD5_H , b, c, d, a, wa_t, MD5C27, MD5S23);
108 MD5_STEP (MD5_H , a, b, c, d, wd_t, MD5C28, MD5S20);
109 MD5_STEP (MD5_H , d, a, b, c, w0_t, MD5C29, MD5S21);
110 MD5_STEP (MD5_H , c, d, a, b, w3_t, MD5C2a, MD5S22);
111 MD5_STEP (MD5_H , b, c, d, a, w6_t, MD5C2b, MD5S23);
112 MD5_STEP (MD5_H , a, b, c, d, w9_t, MD5C2c, MD5S20);
113 MD5_STEP (MD5_H , d, a, b, c, wc_t, MD5C2d, MD5S21);
114 MD5_STEP (MD5_H , c, d, a, b, wf_t, MD5C2e, MD5S22);
115 MD5_STEP (MD5_H , b, c, d, a, w2_t, MD5C2f, MD5S23);
117 MD5_STEP (MD5_I , a, b, c, d, w0_t, MD5C30, MD5S30);
118 MD5_STEP (MD5_I , d, a, b, c, w7_t, MD5C31, MD5S31);
119 MD5_STEP (MD5_I , c, d, a, b, we_t, MD5C32, MD5S32);
120 MD5_STEP (MD5_I , b, c, d, a, w5_t, MD5C33, MD5S33);
121 MD5_STEP (MD5_I , a, b, c, d, wc_t, MD5C34, MD5S30);
122 MD5_STEP (MD5_I , d, a, b, c, w3_t, MD5C35, MD5S31);
123 MD5_STEP (MD5_I , c, d, a, b, wa_t, MD5C36, MD5S32);
124 MD5_STEP (MD5_I , b, c, d, a, w1_t, MD5C37, MD5S33);
125 MD5_STEP (MD5_I , a, b, c, d, w8_t, MD5C38, MD5S30);
126 MD5_STEP (MD5_I , d, a, b, c, wf_t, MD5C39, MD5S31);
127 MD5_STEP (MD5_I , c, d, a, b, w6_t, MD5C3a, MD5S32);
128 MD5_STEP (MD5_I , b, c, d, a, wd_t, MD5C3b, MD5S33);
129 MD5_STEP (MD5_I , a, b, c, d, w4_t, MD5C3c, MD5S30);
130 MD5_STEP (MD5_I , d, a, b, c, wb_t, MD5C3d, MD5S31);
131 MD5_STEP (MD5_I , c, d, a, b, w2_t, MD5C3e, MD5S32);
132 MD5_STEP (MD5_I , b, c, d, a, w9_t, MD5C3f, MD5S33);
140 static void hmac_md5_pad (u32x w0[4], u32x w1[4], u32x w2[4], u32x w3[4], u32x ipad[4], u32x opad[4])
142 w0[0] = w0[0] ^ 0x36363636;
143 w0[1] = w0[1] ^ 0x36363636;
144 w0[2] = w0[2] ^ 0x36363636;
145 w0[3] = w0[3] ^ 0x36363636;
146 w1[0] = w1[0] ^ 0x36363636;
147 w1[1] = w1[1] ^ 0x36363636;
148 w1[2] = w1[2] ^ 0x36363636;
149 w1[3] = w1[3] ^ 0x36363636;
150 w2[0] = w2[0] ^ 0x36363636;
151 w2[1] = w2[1] ^ 0x36363636;
152 w2[2] = w2[2] ^ 0x36363636;
153 w2[3] = w2[3] ^ 0x36363636;
154 w3[0] = w3[0] ^ 0x36363636;
155 w3[1] = w3[1] ^ 0x36363636;
156 w3[2] = w3[2] ^ 0x36363636;
157 w3[3] = w3[3] ^ 0x36363636;
164 md5_transform (w0, w1, w2, w3, ipad);
166 w0[0] = w0[0] ^ 0x6a6a6a6a;
167 w0[1] = w0[1] ^ 0x6a6a6a6a;
168 w0[2] = w0[2] ^ 0x6a6a6a6a;
169 w0[3] = w0[3] ^ 0x6a6a6a6a;
170 w1[0] = w1[0] ^ 0x6a6a6a6a;
171 w1[1] = w1[1] ^ 0x6a6a6a6a;
172 w1[2] = w1[2] ^ 0x6a6a6a6a;
173 w1[3] = w1[3] ^ 0x6a6a6a6a;
174 w2[0] = w2[0] ^ 0x6a6a6a6a;
175 w2[1] = w2[1] ^ 0x6a6a6a6a;
176 w2[2] = w2[2] ^ 0x6a6a6a6a;
177 w2[3] = w2[3] ^ 0x6a6a6a6a;
178 w3[0] = w3[0] ^ 0x6a6a6a6a;
179 w3[1] = w3[1] ^ 0x6a6a6a6a;
180 w3[2] = w3[2] ^ 0x6a6a6a6a;
181 w3[3] = w3[3] ^ 0x6a6a6a6a;
188 md5_transform (w0, w1, w2, w3, opad);
191 static void hmac_md5_run (u32x w0[4], u32x w1[4], u32x w2[4], u32x w3[4], u32x ipad[4], u32x opad[4], u32x digest[4])
198 md5_transform (w0, w1, w2, w3, digest);
214 w3[2] = (64 + 16) * 8;
222 md5_transform (w0, w1, w2, w3, digest);
225 static void m05300m (u32x w0[4], u32x w1[4], u32x w2[4], u32x w3[4], const u32 pw_len, __global pw_t *pws, __global gpu_rule_t *rules_buf, __global comb_t *combs_buf, __global bf_t *bfs_buf, __global void *tmps, __global void *hooks, __global u32 *bitmaps_buf_s1_a, __global u32 *bitmaps_buf_s1_b, __global u32 *bitmaps_buf_s1_c, __global u32 *bitmaps_buf_s1_d, __global u32 *bitmaps_buf_s2_a, __global u32 *bitmaps_buf_s2_b, __global u32 *bitmaps_buf_s2_c, __global u32 *bitmaps_buf_s2_d, __global plain_t *plains_buf, __global digest_t *digests_buf, __global u32 *hashes_shown, __global salt_t *salt_bufs, __global ikepsk_t *ikepsk_bufs, __global u32 *d_return_buf, __global u32 *d_scryptV_buf, const u32 bitmap_mask, const u32 bitmap_shift1, const u32 bitmap_shift2, const u32 salt_pos, const u32 loop_pos, const u32 loop_cnt, const u32 bfs_cnt, const u32 digests_cnt, const u32 digests_offset, __local u32 s_msg_buf[128])
231 const u32 gid = get_global_id (0);
232 const u32 lid = get_local_id (0);
238 const u32 nr_len = ikepsk_bufs[salt_pos].nr_len;
239 const u32 msg_len = ikepsk_bufs[salt_pos].msg_len;
243 salt_buf0[0] = ikepsk_bufs[salt_pos].nr_buf[ 0];
244 salt_buf0[1] = ikepsk_bufs[salt_pos].nr_buf[ 1];
245 salt_buf0[2] = ikepsk_bufs[salt_pos].nr_buf[ 2];
246 salt_buf0[3] = ikepsk_bufs[salt_pos].nr_buf[ 3];
250 salt_buf1[0] = ikepsk_bufs[salt_pos].nr_buf[ 4];
251 salt_buf1[1] = ikepsk_bufs[salt_pos].nr_buf[ 5];
252 salt_buf1[2] = ikepsk_bufs[salt_pos].nr_buf[ 6];
253 salt_buf1[3] = ikepsk_bufs[salt_pos].nr_buf[ 7];
257 salt_buf2[0] = ikepsk_bufs[salt_pos].nr_buf[ 8];
258 salt_buf2[1] = ikepsk_bufs[salt_pos].nr_buf[ 9];
259 salt_buf2[2] = ikepsk_bufs[salt_pos].nr_buf[10];
260 salt_buf2[3] = ikepsk_bufs[salt_pos].nr_buf[11];
264 salt_buf3[0] = ikepsk_bufs[salt_pos].nr_buf[12];
265 salt_buf3[1] = ikepsk_bufs[salt_pos].nr_buf[13];
275 for (u32 il_pos = 0; il_pos < bfs_cnt; il_pos++)
277 const u32 w0r = bfs_buf[il_pos].i;
316 hmac_md5_pad (w0_t, w1_t, w2_t, w3_t, ipad, opad);
318 w0_t[0] = salt_buf0[0];
319 w0_t[1] = salt_buf0[1];
320 w0_t[2] = salt_buf0[2];
321 w0_t[3] = salt_buf0[3];
322 w1_t[0] = salt_buf1[0];
323 w1_t[1] = salt_buf1[1];
324 w1_t[2] = salt_buf1[2];
325 w1_t[3] = salt_buf1[3];
326 w2_t[0] = salt_buf2[0];
327 w2_t[1] = salt_buf2[1];
328 w2_t[2] = salt_buf2[2];
329 w2_t[3] = salt_buf2[3];
330 w3_t[0] = salt_buf3[0];
331 w3_t[1] = salt_buf3[1];
332 w3_t[2] = (64 + nr_len) * 8;
337 hmac_md5_run (w0_t, w1_t, w2_t, w3_t, ipad, opad, digest);
356 hmac_md5_pad (w0_t, w1_t, w2_t, w3_t, ipad, opad);
361 for (left = ikepsk_bufs[salt_pos].msg_len, off = 0; left >= 56; left -= 64, off += 16)
363 w0_t[0] = s_msg_buf[off + 0];
364 w0_t[1] = s_msg_buf[off + 1];
365 w0_t[2] = s_msg_buf[off + 2];
366 w0_t[3] = s_msg_buf[off + 3];
367 w1_t[0] = s_msg_buf[off + 4];
368 w1_t[1] = s_msg_buf[off + 5];
369 w1_t[2] = s_msg_buf[off + 6];
370 w1_t[3] = s_msg_buf[off + 7];
371 w2_t[0] = s_msg_buf[off + 8];
372 w2_t[1] = s_msg_buf[off + 9];
373 w2_t[2] = s_msg_buf[off + 10];
374 w2_t[3] = s_msg_buf[off + 11];
375 w3_t[0] = s_msg_buf[off + 12];
376 w3_t[1] = s_msg_buf[off + 13];
377 w3_t[2] = s_msg_buf[off + 14];
378 w3_t[3] = s_msg_buf[off + 15];
380 md5_transform (w0_t, w1_t, w2_t, w3_t, ipad);
383 w0_t[0] = s_msg_buf[off + 0];
384 w0_t[1] = s_msg_buf[off + 1];
385 w0_t[2] = s_msg_buf[off + 2];
386 w0_t[3] = s_msg_buf[off + 3];
387 w1_t[0] = s_msg_buf[off + 4];
388 w1_t[1] = s_msg_buf[off + 5];
389 w1_t[2] = s_msg_buf[off + 6];
390 w1_t[3] = s_msg_buf[off + 7];
391 w2_t[0] = s_msg_buf[off + 8];
392 w2_t[1] = s_msg_buf[off + 9];
393 w2_t[2] = s_msg_buf[off + 10];
394 w2_t[3] = s_msg_buf[off + 11];
395 w3_t[0] = s_msg_buf[off + 12];
396 w3_t[1] = s_msg_buf[off + 13];
397 w3_t[2] = (64 + msg_len) * 8;
400 hmac_md5_run (w0_t, w1_t, w2_t, w3_t, ipad, opad, digest);
402 const u32x r0 = digest[0];
403 const u32x r1 = digest[3];
404 const u32x r2 = digest[2];
405 const u32x r3 = digest[1];
407 #include VECT_COMPARE_M
411 static void m05300s (u32x w0[4], u32x w1[4], u32x w2[4], u32x w3[4], const u32 pw_len, __global pw_t *pws, __global gpu_rule_t *rules_buf, __global comb_t *combs_buf, __global bf_t *bfs_buf, __global void *tmps, __global void *hooks, __global u32 *bitmaps_buf_s1_a, __global u32 *bitmaps_buf_s1_b, __global u32 *bitmaps_buf_s1_c, __global u32 *bitmaps_buf_s1_d, __global u32 *bitmaps_buf_s2_a, __global u32 *bitmaps_buf_s2_b, __global u32 *bitmaps_buf_s2_c, __global u32 *bitmaps_buf_s2_d, __global plain_t *plains_buf, __global digest_t *digests_buf, __global u32 *hashes_shown, __global salt_t *salt_bufs, __global ikepsk_t *ikepsk_bufs, __global u32 *d_return_buf, __global u32 *d_scryptV_buf, const u32 bitmap_mask, const u32 bitmap_shift1, const u32 bitmap_shift2, const u32 salt_pos, const u32 loop_pos, const u32 loop_cnt, const u32 bfs_cnt, const u32 digests_cnt, const u32 digests_offset, __local u32 s_msg_buf[128])
417 const u32 gid = get_global_id (0);
418 const u32 lid = get_local_id (0);
424 const u32 nr_len = ikepsk_bufs[salt_pos].nr_len;
425 const u32 msg_len = ikepsk_bufs[salt_pos].msg_len;
429 salt_buf0[0] = ikepsk_bufs[salt_pos].nr_buf[ 0];
430 salt_buf0[1] = ikepsk_bufs[salt_pos].nr_buf[ 1];
431 salt_buf0[2] = ikepsk_bufs[salt_pos].nr_buf[ 2];
432 salt_buf0[3] = ikepsk_bufs[salt_pos].nr_buf[ 3];
436 salt_buf1[0] = ikepsk_bufs[salt_pos].nr_buf[ 4];
437 salt_buf1[1] = ikepsk_bufs[salt_pos].nr_buf[ 5];
438 salt_buf1[2] = ikepsk_bufs[salt_pos].nr_buf[ 6];
439 salt_buf1[3] = ikepsk_bufs[salt_pos].nr_buf[ 7];
443 salt_buf2[0] = ikepsk_bufs[salt_pos].nr_buf[ 8];
444 salt_buf2[1] = ikepsk_bufs[salt_pos].nr_buf[ 9];
445 salt_buf2[2] = ikepsk_bufs[salt_pos].nr_buf[10];
446 salt_buf2[3] = ikepsk_bufs[salt_pos].nr_buf[11];
450 salt_buf3[0] = ikepsk_bufs[salt_pos].nr_buf[12];
451 salt_buf3[1] = ikepsk_bufs[salt_pos].nr_buf[13];
459 const u32 search[4] =
461 digests_buf[digests_offset].digest_buf[DGST_R0],
462 digests_buf[digests_offset].digest_buf[DGST_R1],
463 digests_buf[digests_offset].digest_buf[DGST_R2],
464 digests_buf[digests_offset].digest_buf[DGST_R3]
473 for (u32 il_pos = 0; il_pos < bfs_cnt; il_pos++)
475 const u32 w0r = bfs_buf[il_pos].i;
514 hmac_md5_pad (w0_t, w1_t, w2_t, w3_t, ipad, opad);
516 w0_t[0] = salt_buf0[0];
517 w0_t[1] = salt_buf0[1];
518 w0_t[2] = salt_buf0[2];
519 w0_t[3] = salt_buf0[3];
520 w1_t[0] = salt_buf1[0];
521 w1_t[1] = salt_buf1[1];
522 w1_t[2] = salt_buf1[2];
523 w1_t[3] = salt_buf1[3];
524 w2_t[0] = salt_buf2[0];
525 w2_t[1] = salt_buf2[1];
526 w2_t[2] = salt_buf2[2];
527 w2_t[3] = salt_buf2[3];
528 w3_t[0] = salt_buf3[0];
529 w3_t[1] = salt_buf3[1];
530 w3_t[2] = (64 + nr_len) * 8;
535 hmac_md5_run (w0_t, w1_t, w2_t, w3_t, ipad, opad, digest);
554 hmac_md5_pad (w0_t, w1_t, w2_t, w3_t, ipad, opad);
559 for (left = ikepsk_bufs[salt_pos].msg_len, off = 0; left >= 56; left -= 64, off += 16)
561 w0_t[0] = s_msg_buf[off + 0];
562 w0_t[1] = s_msg_buf[off + 1];
563 w0_t[2] = s_msg_buf[off + 2];
564 w0_t[3] = s_msg_buf[off + 3];
565 w1_t[0] = s_msg_buf[off + 4];
566 w1_t[1] = s_msg_buf[off + 5];
567 w1_t[2] = s_msg_buf[off + 6];
568 w1_t[3] = s_msg_buf[off + 7];
569 w2_t[0] = s_msg_buf[off + 8];
570 w2_t[1] = s_msg_buf[off + 9];
571 w2_t[2] = s_msg_buf[off + 10];
572 w2_t[3] = s_msg_buf[off + 11];
573 w3_t[0] = s_msg_buf[off + 12];
574 w3_t[1] = s_msg_buf[off + 13];
575 w3_t[2] = s_msg_buf[off + 14];
576 w3_t[3] = s_msg_buf[off + 15];
578 md5_transform (w0_t, w1_t, w2_t, w3_t, ipad);
581 w0_t[0] = s_msg_buf[off + 0];
582 w0_t[1] = s_msg_buf[off + 1];
583 w0_t[2] = s_msg_buf[off + 2];
584 w0_t[3] = s_msg_buf[off + 3];
585 w1_t[0] = s_msg_buf[off + 4];
586 w1_t[1] = s_msg_buf[off + 5];
587 w1_t[2] = s_msg_buf[off + 6];
588 w1_t[3] = s_msg_buf[off + 7];
589 w2_t[0] = s_msg_buf[off + 8];
590 w2_t[1] = s_msg_buf[off + 9];
591 w2_t[2] = s_msg_buf[off + 10];
592 w2_t[3] = s_msg_buf[off + 11];
593 w3_t[0] = s_msg_buf[off + 12];
594 w3_t[1] = s_msg_buf[off + 13];
595 w3_t[2] = (64 + msg_len) * 8;
598 hmac_md5_run (w0_t, w1_t, w2_t, w3_t, ipad, opad, digest);
600 const u32x r0 = digest[0];
601 const u32x r1 = digest[3];
602 const u32x r2 = digest[2];
603 const u32x r3 = digest[1];
605 #include VECT_COMPARE_S
609 __kernel void __attribute__((reqd_work_group_size (64, 1, 1))) m05300_m04 (__global pw_t *pws, __global gpu_rule_t *rules_buf, __global comb_t *combs_buf, __global bf_t *bfs_buf, __global void *tmps, __global void *hooks, __global u32 *bitmaps_buf_s1_a, __global u32 *bitmaps_buf_s1_b, __global u32 *bitmaps_buf_s1_c, __global u32 *bitmaps_buf_s1_d, __global u32 *bitmaps_buf_s2_a, __global u32 *bitmaps_buf_s2_b, __global u32 *bitmaps_buf_s2_c, __global u32 *bitmaps_buf_s2_d, __global plain_t *plains_buf, __global digest_t *digests_buf, __global u32 *hashes_shown, __global salt_t *salt_bufs, __global ikepsk_t *ikepsk_bufs, __global u32 *d_return_buf, __global u32 *d_scryptV_buf, const u32 bitmap_mask, const u32 bitmap_shift1, const u32 bitmap_shift2, const u32 salt_pos, const u32 loop_pos, const u32 loop_cnt, const u32 bfs_cnt, const u32 digests_cnt, const u32 digests_offset, const u32 combs_mode, const u32 gid_max)
615 const u32 lid = get_local_id (0);
621 const u32 gid = get_global_id (0);
625 w0[0] = pws[gid].i[ 0];
626 w0[1] = pws[gid].i[ 1];
627 w0[2] = pws[gid].i[ 2];
628 w0[3] = pws[gid].i[ 3];
651 const u32 pw_len = pws[gid].pw_len;
657 __local u32 s_msg_buf[128];
659 const u32 lid2 = lid * 2;
661 s_msg_buf[lid2 + 0] = ikepsk_bufs[salt_pos].msg_buf[lid2 + 0];
662 s_msg_buf[lid2 + 1] = ikepsk_bufs[salt_pos].msg_buf[lid2 + 1];
664 barrier (CLK_LOCAL_MEM_FENCE);
666 if (gid >= gid_max) return;
672 m05300m (w0, w1, w2, w3, pw_len, pws, rules_buf, combs_buf, bfs_buf, tmps, hooks, bitmaps_buf_s1_a, bitmaps_buf_s1_b, bitmaps_buf_s1_c, bitmaps_buf_s1_d, bitmaps_buf_s2_a, bitmaps_buf_s2_b, bitmaps_buf_s2_c, bitmaps_buf_s2_d, plains_buf, digests_buf, hashes_shown, salt_bufs, ikepsk_bufs, d_return_buf, d_scryptV_buf, bitmap_mask, bitmap_shift1, bitmap_shift2, salt_pos, loop_pos, loop_cnt, bfs_cnt, digests_cnt, digests_offset, s_msg_buf);
675 __kernel void __attribute__((reqd_work_group_size (64, 1, 1))) m05300_m08 (__global pw_t *pws, __global gpu_rule_t *rules_buf, __global comb_t *combs_buf, __global bf_t *bfs_buf, __global void *tmps, __global void *hooks, __global u32 *bitmaps_buf_s1_a, __global u32 *bitmaps_buf_s1_b, __global u32 *bitmaps_buf_s1_c, __global u32 *bitmaps_buf_s1_d, __global u32 *bitmaps_buf_s2_a, __global u32 *bitmaps_buf_s2_b, __global u32 *bitmaps_buf_s2_c, __global u32 *bitmaps_buf_s2_d, __global plain_t *plains_buf, __global digest_t *digests_buf, __global u32 *hashes_shown, __global salt_t *salt_bufs, __global ikepsk_t *ikepsk_bufs, __global u32 *d_return_buf, __global u32 *d_scryptV_buf, const u32 bitmap_mask, const u32 bitmap_shift1, const u32 bitmap_shift2, const u32 salt_pos, const u32 loop_pos, const u32 loop_cnt, const u32 bfs_cnt, const u32 digests_cnt, const u32 digests_offset, const u32 combs_mode, const u32 gid_max)
681 const u32 lid = get_local_id (0);
687 const u32 gid = get_global_id (0);
691 w0[0] = pws[gid].i[ 0];
692 w0[1] = pws[gid].i[ 1];
693 w0[2] = pws[gid].i[ 2];
694 w0[3] = pws[gid].i[ 3];
698 w1[0] = pws[gid].i[ 4];
699 w1[1] = pws[gid].i[ 5];
700 w1[2] = pws[gid].i[ 6];
701 w1[3] = pws[gid].i[ 7];
717 const u32 pw_len = pws[gid].pw_len;
723 __local u32 s_msg_buf[128];
725 const u32 lid2 = lid * 2;
727 s_msg_buf[lid2 + 0] = ikepsk_bufs[salt_pos].msg_buf[lid2 + 0];
728 s_msg_buf[lid2 + 1] = ikepsk_bufs[salt_pos].msg_buf[lid2 + 1];
730 barrier (CLK_LOCAL_MEM_FENCE);
732 if (gid >= gid_max) return;
738 m05300m (w0, w1, w2, w3, pw_len, pws, rules_buf, combs_buf, bfs_buf, tmps, hooks, bitmaps_buf_s1_a, bitmaps_buf_s1_b, bitmaps_buf_s1_c, bitmaps_buf_s1_d, bitmaps_buf_s2_a, bitmaps_buf_s2_b, bitmaps_buf_s2_c, bitmaps_buf_s2_d, plains_buf, digests_buf, hashes_shown, salt_bufs, ikepsk_bufs, d_return_buf, d_scryptV_buf, bitmap_mask, bitmap_shift1, bitmap_shift2, salt_pos, loop_pos, loop_cnt, bfs_cnt, digests_cnt, digests_offset, s_msg_buf);
741 __kernel void __attribute__((reqd_work_group_size (64, 1, 1))) m05300_m16 (__global pw_t *pws, __global gpu_rule_t *rules_buf, __global comb_t *combs_buf, __global bf_t *bfs_buf, __global void *tmps, __global void *hooks, __global u32 *bitmaps_buf_s1_a, __global u32 *bitmaps_buf_s1_b, __global u32 *bitmaps_buf_s1_c, __global u32 *bitmaps_buf_s1_d, __global u32 *bitmaps_buf_s2_a, __global u32 *bitmaps_buf_s2_b, __global u32 *bitmaps_buf_s2_c, __global u32 *bitmaps_buf_s2_d, __global plain_t *plains_buf, __global digest_t *digests_buf, __global u32 *hashes_shown, __global salt_t *salt_bufs, __global ikepsk_t *ikepsk_bufs, __global u32 *d_return_buf, __global u32 *d_scryptV_buf, const u32 bitmap_mask, const u32 bitmap_shift1, const u32 bitmap_shift2, const u32 salt_pos, const u32 loop_pos, const u32 loop_cnt, const u32 bfs_cnt, const u32 digests_cnt, const u32 digests_offset, const u32 combs_mode, const u32 gid_max)
747 const u32 lid = get_local_id (0);
753 const u32 gid = get_global_id (0);
757 w0[0] = pws[gid].i[ 0];
758 w0[1] = pws[gid].i[ 1];
759 w0[2] = pws[gid].i[ 2];
760 w0[3] = pws[gid].i[ 3];
764 w1[0] = pws[gid].i[ 4];
765 w1[1] = pws[gid].i[ 5];
766 w1[2] = pws[gid].i[ 6];
767 w1[3] = pws[gid].i[ 7];
771 w2[0] = pws[gid].i[ 8];
772 w2[1] = pws[gid].i[ 9];
773 w2[2] = pws[gid].i[10];
774 w2[3] = pws[gid].i[11];
778 w3[0] = pws[gid].i[12];
779 w3[1] = pws[gid].i[13];
783 const u32 pw_len = pws[gid].pw_len;
789 __local u32 s_msg_buf[128];
791 const u32 lid2 = lid * 2;
793 s_msg_buf[lid2 + 0] = ikepsk_bufs[salt_pos].msg_buf[lid2 + 0];
794 s_msg_buf[lid2 + 1] = ikepsk_bufs[salt_pos].msg_buf[lid2 + 1];
796 barrier (CLK_LOCAL_MEM_FENCE);
798 if (gid >= gid_max) return;
804 m05300m (w0, w1, w2, w3, pw_len, pws, rules_buf, combs_buf, bfs_buf, tmps, hooks, bitmaps_buf_s1_a, bitmaps_buf_s1_b, bitmaps_buf_s1_c, bitmaps_buf_s1_d, bitmaps_buf_s2_a, bitmaps_buf_s2_b, bitmaps_buf_s2_c, bitmaps_buf_s2_d, plains_buf, digests_buf, hashes_shown, salt_bufs, ikepsk_bufs, d_return_buf, d_scryptV_buf, bitmap_mask, bitmap_shift1, bitmap_shift2, salt_pos, loop_pos, loop_cnt, bfs_cnt, digests_cnt, digests_offset, s_msg_buf);
807 __kernel void __attribute__((reqd_work_group_size (64, 1, 1))) m05300_s04 (__global pw_t *pws, __global gpu_rule_t *rules_buf, __global comb_t *combs_buf, __global bf_t *bfs_buf, __global void *tmps, __global void *hooks, __global u32 *bitmaps_buf_s1_a, __global u32 *bitmaps_buf_s1_b, __global u32 *bitmaps_buf_s1_c, __global u32 *bitmaps_buf_s1_d, __global u32 *bitmaps_buf_s2_a, __global u32 *bitmaps_buf_s2_b, __global u32 *bitmaps_buf_s2_c, __global u32 *bitmaps_buf_s2_d, __global plain_t *plains_buf, __global digest_t *digests_buf, __global u32 *hashes_shown, __global salt_t *salt_bufs, __global ikepsk_t *ikepsk_bufs, __global u32 *d_return_buf, __global u32 *d_scryptV_buf, const u32 bitmap_mask, const u32 bitmap_shift1, const u32 bitmap_shift2, const u32 salt_pos, const u32 loop_pos, const u32 loop_cnt, const u32 bfs_cnt, const u32 digests_cnt, const u32 digests_offset, const u32 combs_mode, const u32 gid_max)
813 const u32 lid = get_local_id (0);
819 const u32 gid = get_global_id (0);
823 w0[0] = pws[gid].i[ 0];
824 w0[1] = pws[gid].i[ 1];
825 w0[2] = pws[gid].i[ 2];
826 w0[3] = pws[gid].i[ 3];
849 const u32 pw_len = pws[gid].pw_len;
855 __local u32 s_msg_buf[128];
857 const u32 lid2 = lid * 2;
859 s_msg_buf[lid2 + 0] = ikepsk_bufs[salt_pos].msg_buf[lid2 + 0];
860 s_msg_buf[lid2 + 1] = ikepsk_bufs[salt_pos].msg_buf[lid2 + 1];
862 barrier (CLK_LOCAL_MEM_FENCE);
864 if (gid >= gid_max) return;
870 m05300s (w0, w1, w2, w3, pw_len, pws, rules_buf, combs_buf, bfs_buf, tmps, hooks, bitmaps_buf_s1_a, bitmaps_buf_s1_b, bitmaps_buf_s1_c, bitmaps_buf_s1_d, bitmaps_buf_s2_a, bitmaps_buf_s2_b, bitmaps_buf_s2_c, bitmaps_buf_s2_d, plains_buf, digests_buf, hashes_shown, salt_bufs, ikepsk_bufs, d_return_buf, d_scryptV_buf, bitmap_mask, bitmap_shift1, bitmap_shift2, salt_pos, loop_pos, loop_cnt, bfs_cnt, digests_cnt, digests_offset, s_msg_buf);
873 __kernel void __attribute__((reqd_work_group_size (64, 1, 1))) m05300_s08 (__global pw_t *pws, __global gpu_rule_t *rules_buf, __global comb_t *combs_buf, __global bf_t *bfs_buf, __global void *tmps, __global void *hooks, __global u32 *bitmaps_buf_s1_a, __global u32 *bitmaps_buf_s1_b, __global u32 *bitmaps_buf_s1_c, __global u32 *bitmaps_buf_s1_d, __global u32 *bitmaps_buf_s2_a, __global u32 *bitmaps_buf_s2_b, __global u32 *bitmaps_buf_s2_c, __global u32 *bitmaps_buf_s2_d, __global plain_t *plains_buf, __global digest_t *digests_buf, __global u32 *hashes_shown, __global salt_t *salt_bufs, __global ikepsk_t *ikepsk_bufs, __global u32 *d_return_buf, __global u32 *d_scryptV_buf, const u32 bitmap_mask, const u32 bitmap_shift1, const u32 bitmap_shift2, const u32 salt_pos, const u32 loop_pos, const u32 loop_cnt, const u32 bfs_cnt, const u32 digests_cnt, const u32 digests_offset, const u32 combs_mode, const u32 gid_max)
879 const u32 lid = get_local_id (0);
885 const u32 gid = get_global_id (0);
889 w0[0] = pws[gid].i[ 0];
890 w0[1] = pws[gid].i[ 1];
891 w0[2] = pws[gid].i[ 2];
892 w0[3] = pws[gid].i[ 3];
896 w1[0] = pws[gid].i[ 4];
897 w1[1] = pws[gid].i[ 5];
898 w1[2] = pws[gid].i[ 6];
899 w1[3] = pws[gid].i[ 7];
915 const u32 pw_len = pws[gid].pw_len;
921 __local u32 s_msg_buf[128];
923 const u32 lid2 = lid * 2;
925 s_msg_buf[lid2 + 0] = ikepsk_bufs[salt_pos].msg_buf[lid2 + 0];
926 s_msg_buf[lid2 + 1] = ikepsk_bufs[salt_pos].msg_buf[lid2 + 1];
928 barrier (CLK_LOCAL_MEM_FENCE);
930 if (gid >= gid_max) return;
936 m05300s (w0, w1, w2, w3, pw_len, pws, rules_buf, combs_buf, bfs_buf, tmps, hooks, bitmaps_buf_s1_a, bitmaps_buf_s1_b, bitmaps_buf_s1_c, bitmaps_buf_s1_d, bitmaps_buf_s2_a, bitmaps_buf_s2_b, bitmaps_buf_s2_c, bitmaps_buf_s2_d, plains_buf, digests_buf, hashes_shown, salt_bufs, ikepsk_bufs, d_return_buf, d_scryptV_buf, bitmap_mask, bitmap_shift1, bitmap_shift2, salt_pos, loop_pos, loop_cnt, bfs_cnt, digests_cnt, digests_offset, s_msg_buf);
939 __kernel void __attribute__((reqd_work_group_size (64, 1, 1))) m05300_s16 (__global pw_t *pws, __global gpu_rule_t *rules_buf, __global comb_t *combs_buf, __global bf_t *bfs_buf, __global void *tmps, __global void *hooks, __global u32 *bitmaps_buf_s1_a, __global u32 *bitmaps_buf_s1_b, __global u32 *bitmaps_buf_s1_c, __global u32 *bitmaps_buf_s1_d, __global u32 *bitmaps_buf_s2_a, __global u32 *bitmaps_buf_s2_b, __global u32 *bitmaps_buf_s2_c, __global u32 *bitmaps_buf_s2_d, __global plain_t *plains_buf, __global digest_t *digests_buf, __global u32 *hashes_shown, __global salt_t *salt_bufs, __global ikepsk_t *ikepsk_bufs, __global u32 *d_return_buf, __global u32 *d_scryptV_buf, const u32 bitmap_mask, const u32 bitmap_shift1, const u32 bitmap_shift2, const u32 salt_pos, const u32 loop_pos, const u32 loop_cnt, const u32 bfs_cnt, const u32 digests_cnt, const u32 digests_offset, const u32 combs_mode, const u32 gid_max)
945 const u32 lid = get_local_id (0);
951 const u32 gid = get_global_id (0);
955 w0[0] = pws[gid].i[ 0];
956 w0[1] = pws[gid].i[ 1];
957 w0[2] = pws[gid].i[ 2];
958 w0[3] = pws[gid].i[ 3];
962 w1[0] = pws[gid].i[ 4];
963 w1[1] = pws[gid].i[ 5];
964 w1[2] = pws[gid].i[ 6];
965 w1[3] = pws[gid].i[ 7];
969 w2[0] = pws[gid].i[ 8];
970 w2[1] = pws[gid].i[ 9];
971 w2[2] = pws[gid].i[10];
972 w2[3] = pws[gid].i[11];
976 w3[0] = pws[gid].i[12];
977 w3[1] = pws[gid].i[13];
981 const u32 pw_len = pws[gid].pw_len;
987 __local u32 s_msg_buf[128];
989 const u32 lid2 = lid * 2;
991 s_msg_buf[lid2 + 0] = ikepsk_bufs[salt_pos].msg_buf[lid2 + 0];
992 s_msg_buf[lid2 + 1] = ikepsk_bufs[salt_pos].msg_buf[lid2 + 1];
994 barrier (CLK_LOCAL_MEM_FENCE);
996 if (gid >= gid_max) return;
1002 m05300s (w0, w1, w2, w3, pw_len, pws, rules_buf, combs_buf, bfs_buf, tmps, hooks, bitmaps_buf_s1_a, bitmaps_buf_s1_b, bitmaps_buf_s1_c, bitmaps_buf_s1_d, bitmaps_buf_s2_a, bitmaps_buf_s2_b, bitmaps_buf_s2_c, bitmaps_buf_s2_d, plains_buf, digests_buf, hashes_shown, salt_bufs, ikepsk_bufs, d_return_buf, d_scryptV_buf, bitmap_mask, bitmap_shift1, bitmap_shift2, salt_pos, loop_pos, loop_cnt, bfs_cnt, digests_cnt, digests_offset, s_msg_buf);