/**
* Authors.....: Jens Steube <jens.steube@gmail.com>
* Gabriele Gristina <matrix@hashcat.net>
+ * magnum <john.magnum@hushmail.com>
*
* License.....: MIT
*/
#define MARKOV_DISABLE 0
#define MARKOV_CLASSIC 0
#define BENCHMARK 0
+#define BENCHMARK_REPEATS 100
#define RESTORE 0
#define RESTORE_TIMER 60
#define RESTORE_DISABLE 0
#define HL_MODE_FILE 4
#define HL_MODE_ARG 5
+#define HLFMTS_CNT 11
#define HLFMT_HASHCAT 0
#define HLFMT_PWDUMP 1
#define HLFMT_PASSWD 2
#define HLFMT_NETNTLM2 8
#define HLFMT_NSLDAP 9
#define HLFMT_NSLDAPS 10
-#define HLFMTS_CNT 11
+
+#define HLFMT_TEXT_HASHCAT "native hashcat"
+#define HLFMT_TEXT_PWDUMP "pwdump"
+#define HLFMT_TEXT_PASSWD "passwd"
+#define HLFMT_TEXT_SHADOW "shadow"
+#define HLFMT_TEXT_DCC "DCC"
+#define HLFMT_TEXT_DCC2 "DCC 2"
+#define HLFMT_TEXT_NETNTLM1 "NetNTLMv1"
+#define HLFMT_TEXT_NETNTLM2 "NetNTLMv2"
+#define HLFMT_TEXT_NSLDAP "nsldap"
+#define HLFMT_TEXT_NSLDAPS "nsldaps"
#define ATTACK_MODE_STRAIGHT 0
#define ATTACK_MODE_COMBI 1
#define MAX_DICTSTAT 10000
-#define NUM_DEFAULT_BENCHMARK_ALGORITHMS 130
+#define NUM_DEFAULT_BENCHMARK_ALGORITHMS 135
#define global_free(attr) \
{ \
5600,
7300,
7500,
+ 13100,
8300,
11100,
11200,
101,
111,
1711,
- 3000, // broken in osx
+ 3000,
1000,
1100,
2100,
12800,
- 1500, // broken in osx
+ 1500,
12400,
500,
3200,
11600,
12500,
13000,
+ 13200,
+ 13300,
6211,
6221,
6231,
10410,
10500,
10600,
- 10700, // broken in osx
+ 10700,
9000,
5200,
6800,
6600,
8200,
11300,
- 12700
+ 12700,
+ 13400,
+ 125
};
/**
" --remove Enable remove of hash once it is cracked",
" --remove-timer=NUM Update input hash file each NUM seconds",
" --potfile-disable Do not write potfile",
+ " --potfile-path Specific path to potfile",
" --debug-mode=NUM Defines the debug mode (hybrid only by using rules), see references below",
" --debug-file=FILE Output file for debugging rules (see also --debug-mode)",
" --induction-dir=FOLDER Specify induction directory to use, default is $session.induct",
"* Resources:",
"",
" -b, --benchmark Run benchmark",
+ " --benchmark-repeats=NUM Repeat the kernel on the device NUM times to increase benchmark accuracy",
" -c, --segment-size=NUM Size in MB to cache from the wordfile",
" --bitmap-min=NUM Minimum number of bits allowed for bitmaps",
" --bitmap-max=NUM Maximum number of bits allowed for bitmaps",
" --opencl-platforms=STR OpenCL platforms to use, separate with comma",
" -d, --opencl-devices=STR OpenCL devices to use, separate with comma",
" --opencl-device-types=STR OpenCL device-types to use, separate with comma, see references below",
- " --opencl-vector-width=NUM OpenCL vector-width (either 1, 2, 4 or 8), overrides value from device query",
+ " --opencl-vector-width=NUM OpenCL vector-width (either 1, 2, 4, 8 or 16), overrides value from device query",
" -w, --workload-profile=NUM Enable a specific workload profile, see references below",
" -n, --kernel-accel=NUM Workload tuning, increase the outer-loop step size",
" -u, --kernel-loops=NUM Workload tuning, increase the inner-loop step size",
" 11100 = PostgreSQL Challenge-Response Authentication (MD5)",
" 11200 = MySQL Challenge-Response Authentication (SHA1)",
" 11400 = SIP digest authentication (MD5)",
+ " 13100 = Kerberos 5 TGS-REP etype 23",
"",
"[[ Forums, CMS, E-Commerce, Frameworks, Middleware, Wiki, Management ]]",
"",
" 8500 = RACF",
" 7200 = GRUB 2",
" 9900 = Radmin2",
+ " 125 = ArubaOS",
"",
"[[ Enterprise Application Software (EAS) ]]",
"",
" 11600 = 7-Zip",
" 12500 = RAR3-hp",
" 13000 = RAR5",
+ " 13200 = AxCrypt",
+ " 13300 = AxCrypt in memory SHA1",
"",
"[[ Full-Disk encryptions (FDE) ]]",
"",
" 8200 = 1Password, cloudkeychain",
" 11300 = Bitcoin/Litecoin wallet.dat",
" 12700 = Blockchain, My Wallet",
+ " 13400 = Keepass 1 (AES/Twofish) and Keepass 2 (AES)",
"",
NULL
};
if (device_param->skipped) continue;
- u64 speed_cnt = 0;
- float speed_ms = 0;
+ u64 speed_cnt = 0;
+ double speed_ms = 0;
for (int i = 0; i < SPEED_CACHE; i++)
{
- float rec_ms;
-
- hc_timer_get (device_param->speed_rec[i], rec_ms);
-
- if (rec_ms > SPEED_MAXAGE) continue;
-
speed_cnt += device_param->speed_cnt[i];
speed_ms += device_param->speed_ms[i];
}
* counter
*/
- uint salts_left = data.salts_cnt - data.salts_done;
-
- if (salts_left == 0) salts_left = 1;
-
- u64 progress_total = data.words_cnt * salts_left;
+ u64 progress_total = data.words_cnt * data.salts_cnt;
u64 all_done = 0;
u64 all_rejected = 0;
for (uint salt_pos = 0; salt_pos < data.salts_cnt; salt_pos++)
{
- if (salts_left > 1)
- {
- // otherwise the final cracked status shows 0/XXX progress
-
- if (data.salts_shown[salt_pos] == 1) continue;
- }
-
all_done += data.words_progress_done[salt_pos];
all_rejected += data.words_progress_rejected[salt_pos];
all_restored += data.words_progress_restored[salt_pos];
if (data.skip)
{
- progress_skip = MIN (data.skip, data.words_base) * salts_left;
+ progress_skip = MIN (data.skip, data.words_base) * data.salts_cnt;
if (data.attack_kern == ATTACK_KERN_STRAIGHT) progress_skip *= data.kernel_rules_cnt;
else if (data.attack_kern == ATTACK_KERN_COMBI) progress_skip *= data.combs_cnt;
if (data.limit)
{
- progress_end = MIN (data.limit, data.words_base) * salts_left;
+ progress_end = MIN (data.limit, data.words_base) * data.salts_cnt;
if (data.attack_kern == ATTACK_KERN_STRAIGHT) progress_end *= data.kernel_rules_cnt;
else if (data.attack_kern == ATTACK_KERN_COMBI) progress_end *= data.combs_cnt;
{
wpa_t *wpa = (wpa_t *) data.esalts_buf;
- uint pke[25] = { 0 };
-
- char *pke_ptr = (char *) pke;
-
- for (uint i = 0; i < 25; i++)
- {
- pke[i] = byte_swap_32 (wpa->pke[i]);
- }
-
- char mac1[6] = { 0 };
- char mac2[6] = { 0 };
-
- memcpy (mac1, pke_ptr + 23, 6);
- memcpy (mac2, pke_ptr + 29, 6);
-
log_info ("Hash.Target....: %s (%02x:%02x:%02x:%02x:%02x:%02x <-> %02x:%02x:%02x:%02x:%02x:%02x)",
(char *) data.salts_buf[0].salt_buf,
- mac1[0] & 0xff,
- mac1[1] & 0xff,
- mac1[2] & 0xff,
- mac1[3] & 0xff,
- mac1[4] & 0xff,
- mac1[5] & 0xff,
- mac2[0] & 0xff,
- mac2[1] & 0xff,
- mac2[2] & 0xff,
- mac2[3] & 0xff,
- mac2[4] & 0xff,
- mac2[5] & 0xff);
+ wpa->orig_mac1[0],
+ wpa->orig_mac1[1],
+ wpa->orig_mac1[2],
+ wpa->orig_mac1[3],
+ wpa->orig_mac1[4],
+ wpa->orig_mac1[5],
+ wpa->orig_mac2[0],
+ wpa->orig_mac2[1],
+ wpa->orig_mac2[2],
+ wpa->orig_mac2[3],
+ wpa->orig_mac2[4],
+ wpa->orig_mac2[5]);
}
else if (data.hash_mode == 5200)
{
}
else
{
- char out_buf[4096] = { 0 };
+ char out_buf[HCBUFSIZ] = { 0 };
ascii_digest (out_buf, 0, 0);
{
if (data.hash_mode == 3000)
{
- char out_buf1[4096] = { 0 };
- char out_buf2[4096] = { 0 };
+ char out_buf1[32] = { 0 };
+ char out_buf2[32] = { 0 };
ascii_digest (out_buf1, 0, 0);
ascii_digest (out_buf2, 0, 1);
* speed new
*/
- u64 speed_cnt[DEVICES_MAX] = { 0 };
- float speed_ms[DEVICES_MAX] = { 0 };
+ u64 speed_cnt[DEVICES_MAX] = { 0 };
+ double speed_ms[DEVICES_MAX] = { 0 };
for (uint device_id = 0; device_id < data.devices_cnt; device_id++)
{
if (device_param->skipped) continue;
- // we need to clear values (set to 0) because in case the device does
- // not get new candidates it idles around but speed display would
- // show it as working.
- // if we instantly set it to 0 after reading it happens that the
- // speed can be shown as zero if the users refreshes too fast.
- // therefore, we add a timestamp when a stat was recorded and if its
- // too old we will not use it
-
speed_cnt[device_id] = 0;
speed_ms[device_id] = 0;
for (int i = 0; i < SPEED_CACHE; i++)
{
- float rec_ms;
-
- hc_timer_get (device_param->speed_rec[i], rec_ms);
-
- if (rec_ms > SPEED_MAXAGE) continue;
-
speed_cnt[device_id] += device_param->speed_cnt[i];
speed_ms[device_id] += device_param->speed_ms[i];
}
* timers
*/
- float ms_running = 0;
+ double ms_running = 0;
hc_timer_get (data.timer_running, ms_running);
- float ms_paused = data.ms_paused;
+ double ms_paused = data.ms_paused;
if (data.devices_status == STATUS_PAUSED)
{
- float ms_paused_tmp = 0;
+ double ms_paused_tmp = 0;
hc_timer_get (data.timer_paused, ms_paused_tmp);
* counters
*/
- uint salts_left = data.salts_cnt - data.salts_done;
-
- if (salts_left == 0) salts_left = 1;
-
- u64 progress_total = data.words_cnt * salts_left;
+ u64 progress_total = data.words_cnt * data.salts_cnt;
u64 all_done = 0;
u64 all_rejected = 0;
u64 all_restored = 0;
+ u64 progress_noneed = 0;
+
for (uint salt_pos = 0; salt_pos < data.salts_cnt; salt_pos++)
{
- if (salts_left > 1)
- {
- // otherwise the final cracked status shows 0/XXX progress
-
- if (data.salts_shown[salt_pos] == 1) continue;
- }
-
all_done += data.words_progress_done[salt_pos];
all_rejected += data.words_progress_rejected[salt_pos];
all_restored += data.words_progress_restored[salt_pos];
+
+ // Important for ETA only
+
+ if (data.salts_shown[salt_pos] == 1)
+ {
+ const u64 all = data.words_progress_done[salt_pos]
+ + data.words_progress_rejected[salt_pos]
+ + data.words_progress_restored[salt_pos];
+
+ const u64 left = data.words_cnt - all;
+
+ progress_noneed += left;
+ }
}
u64 progress_cur = all_restored + all_done + all_rejected;
if (data.skip)
{
- progress_skip = MIN (data.skip, data.words_base) * salts_left;
+ progress_skip = MIN (data.skip, data.words_base) * data.salts_cnt;
if (data.attack_kern == ATTACK_KERN_STRAIGHT) progress_skip *= data.kernel_rules_cnt;
else if (data.attack_kern == ATTACK_KERN_COMBI) progress_skip *= data.combs_cnt;
if (data.limit)
{
- progress_end = MIN (data.limit, data.words_base) * salts_left;
+ progress_end = MIN (data.limit, data.words_base) * data.salts_cnt;
if (data.attack_kern == ATTACK_KERN_STRAIGHT) progress_end *= data.kernel_rules_cnt;
else if (data.attack_kern == ATTACK_KERN_COMBI) progress_end *= data.combs_cnt;
u64 progress_cur_relative_skip = progress_cur - progress_skip;
u64 progress_end_relative_skip = progress_end - progress_skip;
- float speed_ms_real = ms_running - ms_paused;
- u64 speed_plains_real = all_done;
-
if ((data.wordlist_mode == WL_MODE_FILE) || (data.wordlist_mode == WL_MODE_MASK))
{
if (data.devices_status != STATUS_CRACKED)
{
- u64 words_per_ms = 0;
-
- if (speed_plains_real && speed_ms_real)
- {
- words_per_ms = speed_plains_real / speed_ms_real;
- }
-
#ifdef WIN
__time64_t sec_etc = 0;
#else
time_t sec_etc = 0;
#endif
- if (words_per_ms)
+ if (hashes_all_ms)
{
u64 progress_left_relative_skip = progress_end_relative_skip - progress_cur_relative_skip;
- u64 ms_left = progress_left_relative_skip / words_per_ms;
+ u64 ms_left = (progress_left_relative_skip - progress_noneed) / hashes_all_ms;
sec_etc = ms_left / 1000;
}
if (sec_etc == 0)
{
- log_info ("Time.Estimated.: 0 secs");
+ //log_info ("Time.Estimated.: 0 secs");
}
else if ((u64) sec_etc > ETC_MAX)
{
}
}
- float cpt_avg_min = (float) data.cpt_total / ((speed_ms_real / 1000) / 60);
- float cpt_avg_hour = (float) data.cpt_total / ((speed_ms_real / 1000) / 3600);
- float cpt_avg_day = (float) data.cpt_total / ((speed_ms_real / 1000) / 86400);
+ double ms_real = ms_running - ms_paused;
+
+ float cpt_avg_min = (float) data.cpt_total / ((ms_real / 1000) / 60);
+ float cpt_avg_hour = (float) data.cpt_total / ((ms_real / 1000) / 3600);
+ float cpt_avg_day = (float) data.cpt_total / ((ms_real / 1000) / 86400);
if ((data.cpt_start + 86400) < now)
{
static void status_benchmark ()
{
- if (data.devices_status == STATUS_INIT) return;
+ if (data.devices_status == STATUS_INIT) return;
if (data.devices_status == STATUS_STARTING) return;
if (data.words_cnt == 0) return;
- u64 speed_cnt[DEVICES_MAX] = { 0 };
- float speed_ms[DEVICES_MAX] = { 0 };
+ u64 speed_cnt[DEVICES_MAX] = { 0 };
+ double speed_ms[DEVICES_MAX] = { 0 };
for (uint device_id = 0; device_id < data.devices_cnt; device_id++)
{
if (device_param->skipped) continue;
- speed_cnt[device_id] = 0;
- speed_ms[device_id] = 0;
-
- for (int i = 0; i < SPEED_CACHE; i++)
- {
- speed_cnt[device_id] += device_param->speed_cnt[i];
- speed_ms[device_id] += device_param->speed_ms[i];
- }
-
- speed_cnt[device_id] /= SPEED_CACHE;
- speed_ms[device_id] /= SPEED_CACHE;
+ speed_cnt[device_id] = device_param->speed_cnt[0];
+ speed_ms[device_id] = device_param->speed_ms[0];
}
float hashes_all_ms = 0;
// hash
- char out_buf[4096] = { 0 };
+ char out_buf[HCBUFSIZ] = { 0 };
ascii_digest (out_buf, salt_pos, digest_pos);
for (int i = 0, j = gidm; i < 16; i++, j++)
{
- plain_buf[i] = pw.h.hi1[0][j];
+ plain_buf[i] = pw.i[j];
}
plain_len = pw.pw_len;
for (int i = 0, j = gidm; i < 16; i++, j++)
{
- plain_buf[i] = pw.h.hi1[0][j];
+ plain_buf[i] = pw.i[j];
}
plain_len = pw.pw_len;
for (int i = 0, j = gidm; i < 16; i++, j++)
{
- plain_buf[i] = pw.h.hi1[0][j];
+ plain_buf[i] = pw.i[j];
}
plain_len = pw.pw_len;
for (int i = 0, j = gidm; i < 16; i++, j++)
{
- plain_buf[i] = pw.h.hi1[0][j];
+ plain_buf[i] = pw.i[j];
}
plain_len = pw.pw_len;
if (data.hash_mode != 2500)
{
- char out_buf[4096] = { 0 };
+ char out_buf[HCBUFSIZ] = { 0 };
if (data.username == 1)
{
hc_clSetKernelArg (data.ocl, kernel, 30, sizeof (cl_uint), device_param->kernel_params[30]);
hc_clSetKernelArg (data.ocl, kernel, 31, sizeof (cl_uint), device_param->kernel_params[31]);
- cl_event event;
+ hc_timer_t timer;
+
+ hc_timer_set (&timer);
if ((data.opts_type & OPTS_TYPE_PT_BITSLICE) && (data.attack_mode == ATTACK_MODE_BF))
{
const size_t global_work_size[3] = { num_elements, 32, 1 };
const size_t local_work_size[3] = { kernel_threads / 32, 32, 1 };
- hc_clEnqueueNDRangeKernel (data.ocl, device_param->command_queue, kernel, 2, NULL, global_work_size, local_work_size, 0, NULL, &event, true);
+ hc_clEnqueueNDRangeKernel (data.ocl, device_param->command_queue, kernel, 2, NULL, global_work_size, local_work_size, 0, NULL, NULL);
}
else
{
- const size_t global_work_size[3] = { num_elements, 1, 1 };
- const size_t local_work_size[3] = { kernel_threads, 1, 1 };
+ size_t workgroup_size = 0;
- const cl_int rc = hc_clEnqueueNDRangeKernel (data.ocl, device_param->command_queue, kernel, 1, NULL, global_work_size, local_work_size, 0, NULL, &event, false);
+ hc_clGetKernelWorkGroupInfo (data.ocl, kernel, device_param->device, CL_KERNEL_WORK_GROUP_SIZE, sizeof (size_t), &workgroup_size, NULL);
- if (rc != CL_SUCCESS)
- {
- const size_t local_work_size_fallback[3] = { 1, 1, 1 };
+ if (kernel_threads > workgroup_size) kernel_threads = workgroup_size;
- hc_clEnqueueNDRangeKernel (data.ocl, device_param->command_queue, kernel, 1, NULL, global_work_size, local_work_size_fallback, 0, NULL, &event, true);
- }
+ const size_t global_work_size[3] = { num_elements, 1, 1 };
+ const size_t local_work_size[3] = { kernel_threads, 1, 1 };
+
+ hc_clEnqueueNDRangeKernel (data.ocl, device_param->command_queue, kernel, 1, NULL, global_work_size, local_work_size, 0, NULL, NULL);
}
hc_clFlush (data.ocl, device_param->command_queue);
- hc_clWaitForEvents (data.ocl, 1, &event);
+ hc_clFinish (data.ocl, device_param->command_queue);
if (event_update)
{
- cl_ulong time_start;
- cl_ulong time_end;
+ double exec_time;
- hc_clGetEventProfilingInfo (data.ocl, event, CL_PROFILING_COMMAND_START, sizeof (time_start), &time_start, NULL);
- hc_clGetEventProfilingInfo (data.ocl, event, CL_PROFILING_COMMAND_END, sizeof (time_end), &time_end, NULL);
-
- const double exec_time = (time_end - time_start) / 1000000.0;
+ hc_timer_get (timer, exec_time);
uint exec_pos = device_param->exec_pos;
device_param->exec_pos = exec_pos;
}
-
- hc_clReleaseEvent (data.ocl, event);
-
- hc_clFinish (data.ocl, device_param->command_queue);
}
static void run_kernel_mp (const uint kern_run, hc_device_param_t *device_param, const uint num)
// causes problems with special threads like in bcrypt
// const uint kernel_threads = device_param->kernel_threads;
- const uint kernel_threads = KERNEL_THREADS;
+ uint kernel_threads = KERNEL_THREADS;
while (num_elements % kernel_threads) num_elements++;
break;
}
- const size_t global_work_size[3] = { num_elements, 1, 1 };
- const size_t local_work_size[3] = { kernel_threads, 1, 1 };
-
- const cl_int rc = hc_clEnqueueNDRangeKernel (data.ocl, device_param->command_queue, kernel, 1, NULL, global_work_size, local_work_size, 0, NULL, NULL, false);
-
- if (rc != CL_SUCCESS)
- {
- const size_t local_work_size_fallback[3] = { 1, 1, 1 };
-
- hc_clEnqueueNDRangeKernel (data.ocl, device_param->command_queue, kernel, 1, NULL, global_work_size, local_work_size_fallback, 0, NULL, NULL, true);
- }
-
- hc_clFlush (data.ocl, device_param->command_queue);
-
- hc_clFinish (data.ocl, device_param->command_queue);
-}
-
-static void run_kernel_tb (hc_device_param_t *device_param, const uint num)
-{
- uint num_elements = num;
-
- uint kernel_threads = device_param->kernel_threads;
-
- while (num_elements % kernel_threads) num_elements++;
-
- cl_kernel kernel = device_param->kernel_tb;
+ size_t workgroup_size = 0;
+ hc_clGetKernelWorkGroupInfo (data.ocl, kernel, device_param->device, CL_KERNEL_WORK_GROUP_SIZE, sizeof(size_t), &workgroup_size, NULL);
+ if (kernel_threads > workgroup_size) kernel_threads = workgroup_size;
const size_t global_work_size[3] = { num_elements, 1, 1 };
- const size_t local_work_size[3] = { kernel_threads, 1, 1 };
-
- const cl_int rc = hc_clEnqueueNDRangeKernel (data.ocl, device_param->command_queue, kernel, 1, NULL, global_work_size, local_work_size, 0, NULL, NULL, false);
+ const size_t local_work_size[3] = { kernel_threads, 1, 1 };
- if (rc != CL_SUCCESS)
- {
- const size_t local_work_size_fallback[3] = { 1, 1, 1 };
-
- hc_clEnqueueNDRangeKernel (data.ocl, device_param->command_queue, kernel, 1, NULL, global_work_size, local_work_size_fallback, 0, NULL, NULL, true);
- }
+ hc_clEnqueueNDRangeKernel (data.ocl, device_param->command_queue, kernel, 1, NULL, global_work_size, local_work_size, 0, NULL, NULL);
hc_clFlush (data.ocl, device_param->command_queue);
{
const uint num_elements = 1024; // fixed
- const uint kernel_threads = 32;
+ uint kernel_threads = 32;
cl_kernel kernel = device_param->kernel_tm;
+ size_t workgroup_size = 0;
+ hc_clGetKernelWorkGroupInfo (data.ocl, kernel, device_param->device, CL_KERNEL_WORK_GROUP_SIZE, sizeof (size_t), &workgroup_size, NULL);
+ if (kernel_threads > workgroup_size) kernel_threads = workgroup_size;
+
const size_t global_work_size[3] = { num_elements, 1, 1 };
const size_t local_work_size[3] = { kernel_threads, 1, 1 };
- const cl_int rc = hc_clEnqueueNDRangeKernel (data.ocl, device_param->command_queue, kernel, 1, NULL, global_work_size, local_work_size, 0, NULL, NULL, false);
-
- if (rc != CL_SUCCESS)
- {
- const size_t local_work_size_fallback[3] = { 1, 1, 1 };
-
- hc_clEnqueueNDRangeKernel (data.ocl, device_param->command_queue, kernel, 1, NULL, global_work_size, local_work_size_fallback, 0, NULL, NULL, true);
- }
+ hc_clEnqueueNDRangeKernel (data.ocl, device_param->command_queue, kernel, 1, NULL, global_work_size, local_work_size, 0, NULL, NULL);
hc_clFlush (data.ocl, device_param->command_queue);
// causes problems with special threads like in bcrypt
// const uint kernel_threads = device_param->kernel_threads;
- const uint kernel_threads = KERNEL_THREADS;
+ uint kernel_threads = KERNEL_THREADS;
while (num_elements % kernel_threads) num_elements++;
hc_clSetKernelArg (data.ocl, kernel, 5, sizeof (cl_uint), device_param->kernel_params_amp[5]);
hc_clSetKernelArg (data.ocl, kernel, 6, sizeof (cl_uint), device_param->kernel_params_amp[6]);
+ size_t workgroup_size = 0;
+ hc_clGetKernelWorkGroupInfo (data.ocl, kernel, device_param->device, CL_KERNEL_WORK_GROUP_SIZE, sizeof (size_t), &workgroup_size, NULL);
+ if (kernel_threads > workgroup_size) kernel_threads = workgroup_size;
+
const size_t global_work_size[3] = { num_elements, 1, 1 };
const size_t local_work_size[3] = { kernel_threads, 1, 1 };
- const cl_int rc = hc_clEnqueueNDRangeKernel (data.ocl, device_param->command_queue, kernel, 1, NULL, global_work_size, local_work_size, 0, NULL, NULL, false);
-
- if (rc != CL_SUCCESS)
- {
- const size_t local_work_size_fallback[3] = { 1, 1, 1 };
-
- hc_clEnqueueNDRangeKernel (data.ocl, device_param->command_queue, kernel, 1, NULL, global_work_size, local_work_size_fallback, 0, NULL, NULL, true);
- }
+ hc_clEnqueueNDRangeKernel (data.ocl, device_param->command_queue, kernel, 1, NULL, global_work_size, local_work_size, 0, NULL, NULL);
hc_clFlush (data.ocl, device_param->command_queue);
}
}
+static void choose_kernel (hc_device_param_t *device_param, const uint attack_exec, const uint attack_mode, const uint opts_type, const salt_t *salt_buf, const uint highest_pw_len, const uint pws_cnt)
+{
+ if (attack_exec == ATTACK_EXEC_INSIDE_KERNEL)
+ {
+ if (attack_mode == ATTACK_MODE_BF)
+ {
+ if (opts_type & OPTS_TYPE_PT_BITSLICE)
+ {
+ const uint size_tm = 32 * sizeof (bs_word_t);
+
+ run_kernel_bzero (device_param, device_param->d_tm_c, size_tm);
+
+ run_kernel_tm (device_param);
+
+ hc_clEnqueueCopyBuffer (data.ocl, device_param->command_queue, device_param->d_tm_c, device_param->d_bfs_c, 0, 0, size_tm, 0, NULL, NULL);
+ }
+ }
+
+ if (highest_pw_len < 16)
+ {
+ run_kernel (KERN_RUN_1, device_param, pws_cnt, true);
+ }
+ else if (highest_pw_len < 32)
+ {
+ run_kernel (KERN_RUN_2, device_param, pws_cnt, true);
+ }
+ else
+ {
+ run_kernel (KERN_RUN_3, device_param, pws_cnt, true);
+ }
+ }
+ else
+ {
+ run_kernel_amp (device_param, pws_cnt);
+
+ run_kernel (KERN_RUN_1, device_param, pws_cnt, false);
+
+ if (opts_type & OPTS_TYPE_HOOK12)
+ {
+ run_kernel (KERN_RUN_12, device_param, pws_cnt, false);
+ }
+
+ uint iter = salt_buf->salt_iter;
+
+ uint loop_step = device_param->kernel_loops;
+
+ for (uint loop_pos = 0; loop_pos < iter; loop_pos += loop_step)
+ {
+ uint loop_left = iter - loop_pos;
+
+ loop_left = MIN (loop_left, loop_step);
+
+ device_param->kernel_params_buf32[25] = loop_pos;
+ device_param->kernel_params_buf32[26] = loop_left;
+
+ run_kernel (KERN_RUN_2, device_param, pws_cnt, true);
+
+ if (data.devices_status == STATUS_CRACKED) break;
+ if (data.devices_status == STATUS_ABORTED) break;
+ if (data.devices_status == STATUS_QUIT) break;
+
+ /**
+ * speed
+ */
+
+ const float iter_part = (float) (loop_pos + loop_left) / iter;
+
+ const u64 perf_sum_all = pws_cnt * iter_part;
+
+ double speed_ms;
+
+ hc_timer_get (device_param->timer_speed, speed_ms);
+
+ const u32 speed_pos = device_param->speed_pos;
+
+ device_param->speed_cnt[speed_pos] = perf_sum_all;
+
+ device_param->speed_ms[speed_pos] = speed_ms;
+ }
+
+ if (opts_type & OPTS_TYPE_HOOK23)
+ {
+ run_kernel (KERN_RUN_23, device_param, pws_cnt, false);
+
+ hc_clEnqueueReadBuffer (data.ocl, device_param->command_queue, device_param->d_hooks, CL_TRUE, 0, device_param->size_hooks, device_param->hooks_buf, 0, NULL, NULL);
+
+ // do something with data
+
+ hc_clEnqueueWriteBuffer (data.ocl, device_param->command_queue, device_param->d_hooks, CL_TRUE, 0, device_param->size_hooks, device_param->hooks_buf, 0, NULL, NULL);
+ }
+
+ run_kernel (KERN_RUN_3, device_param, pws_cnt, false);
+ }
+}
+
static int run_rule_engine (const int rule_len, const char *rule_buf)
{
if (rule_len == 0)
}
else if (data.attack_kern == ATTACK_KERN_COMBI)
{
+ if (data.attack_mode == ATTACK_MODE_HYBRID2)
+ {
+ if (data.opts_type & OPTS_TYPE_PT_ADD01)
+ {
+ for (u32 i = 0; i < pws_cnt; i++)
+ {
+ const u32 pw_len = device_param->pws_buf[i].pw_len;
+
+ u8 *ptr = (u8 *) device_param->pws_buf[i].i;
+
+ ptr[pw_len] = 0x01;
+ }
+ }
+ else if (data.opts_type & OPTS_TYPE_PT_ADD80)
+ {
+ for (u32 i = 0; i < pws_cnt; i++)
+ {
+ const u32 pw_len = device_param->pws_buf[i].pw_len;
+
+ u8 *ptr = (u8 *) device_param->pws_buf[i].i;
+
+ ptr[pw_len] = 0x80;
+ }
+ }
+ }
+
hc_clEnqueueWriteBuffer (data.ocl, device_param->command_queue, device_param->d_pws_buf, CL_TRUE, 0, pws_cnt * sizeof (pw_t), device_param->pws_buf, 0, NULL, NULL);
}
else if (data.attack_kern == ATTACK_KERN_BF)
device_param->kernel_params_buf32[26] = kernel_loops;
device_param->kernel_params_buf32[27] = kernel_loops;
+ // init some fake words
+
+ for (u32 i = 0; i < kernel_power; i++)
+ {
+ device_param->pws_buf[i].i[0] = i;
+ device_param->pws_buf[i].i[1] = 0x01234567;
+ device_param->pws_buf[i].pw_len = 4 + (i & 3);
+ }
+
+ hc_clEnqueueWriteBuffer (data.ocl, device_param->command_queue, device_param->d_pws_buf, CL_TRUE, 0, kernel_power * sizeof (pw_t), device_param->pws_buf, 0, NULL, NULL);
+
+ if (data.attack_exec == ATTACK_EXEC_OUTSIDE_KERNEL)
+ {
+ run_kernel_amp (device_param, kernel_power);
+ }
+
// caching run
if (data.attack_exec == ATTACK_EXEC_INSIDE_KERNEL)
const double exec_ms_prev = get_avg_exec_time (device_param, repeat);
+ // reset fake words
+
+ memset (device_param->pws_buf, 0, kernel_power * sizeof (pw_t));
+
+ hc_clEnqueueWriteBuffer (data.ocl, device_param->command_queue, device_param->d_pws_buf, CL_TRUE, 0, kernel_power * sizeof (pw_t), device_param->pws_buf, 0, NULL, NULL);
+ hc_clEnqueueWriteBuffer (data.ocl, device_param->command_queue, device_param->d_pws_amp_buf, CL_TRUE, 0, kernel_power * sizeof (pw_t), device_param->pws_buf, 0, NULL, NULL);
+
return exec_ms_prev;
}
{
const double target_ms = TARGET_MS_PROFILE[data.workload_profile - 1];
- u32 kernel_loops_min = device_param->kernel_loops_min;
- u32 kernel_loops_max = device_param->kernel_loops_max;
+ const u32 kernel_accel_min = device_param->kernel_accel_min;
+ const u32 kernel_accel_max = device_param->kernel_accel_max;
- u32 kernel_accel_min = device_param->kernel_accel_min;
- u32 kernel_accel_max = device_param->kernel_accel_max;
+ const u32 kernel_loops_min = device_param->kernel_loops_min;
+ const u32 kernel_loops_max = device_param->kernel_loops_max;
- u32 kernel_loops = kernel_loops_min;
u32 kernel_accel = kernel_accel_min;
+ u32 kernel_loops = kernel_loops_min;
- // init some fake words
-
- const u32 kernel_power_max = device_param->device_processors * device_param->kernel_threads * device_param->kernel_accel_max;
+ // steps
- for (u32 i = 0; i < kernel_power_max; i++)
- {
- device_param->pws_buf[i].pw_len = 8;
- }
+ #define STEPS_CNT 10
- hc_clEnqueueWriteBuffer (data.ocl, device_param->command_queue, device_param->d_pws_buf, CL_TRUE, 0, device_param->size_pws, device_param->pws_buf, 0, NULL, NULL);
- hc_clEnqueueWriteBuffer (data.ocl, device_param->command_queue, device_param->d_pws_amp_buf, CL_TRUE, 0, device_param->size_pws, device_param->pws_buf, 0, NULL, NULL);
+ #define STEPS_ACCEL_CNT (STEPS_CNT + 2)
+ #define STEPS_LOOPS_CNT (STEPS_CNT + 2)
- // steps for loops
+ u32 steps_accel[STEPS_ACCEL_CNT];
+ u32 steps_loops[STEPS_LOOPS_CNT];
- #define STEPS_LOOPS_CNT 15
+ for (int i = 0; i < STEPS_ACCEL_CNT; i++)
+ {
+ steps_accel[i] = 1 << i;
+ }
- u32 steps_loops[STEPS_LOOPS_CNT];
+ for (int i = 0; i < STEPS_LOOPS_CNT; i++)
+ {
+ steps_loops[i] = 1 << i;
+ }
- steps_loops[ 0] = 1;
- steps_loops[ 1] = 2;
- steps_loops[ 2] = 4;
- steps_loops[ 3] = 8;
- steps_loops[ 4] = 16;
- steps_loops[ 5] = 32;
- steps_loops[ 6] = 64;
- steps_loops[ 7] = 100;
- steps_loops[ 8] = 128;
- steps_loops[ 9] = 200;
- steps_loops[10] = 256;
- steps_loops[11] = 500;
- steps_loops[12] = 512;
- steps_loops[13] = 1000;
- steps_loops[14] = 1024;
-
- // steps for accel
-
- #define STEPS_ACCEL_CNT 13
+ steps_accel[STEPS_CNT + 0] = kernel_accel_min;
+ steps_accel[STEPS_CNT + 1] = kernel_accel_max;
- u32 steps_accel[STEPS_ACCEL_CNT];
+ steps_loops[STEPS_CNT + 0] = kernel_loops_min;
+ steps_loops[STEPS_CNT + 1] = kernel_loops_max;
- steps_accel[ 0] = 1;
- steps_accel[ 1] = 2;
- steps_accel[ 2] = 4;
- steps_accel[ 3] = 8;
- steps_accel[ 4] = 16;
- steps_accel[ 5] = 32;
- steps_accel[ 6] = 64;
- steps_accel[ 7] = 128;
- steps_accel[ 8] = 256;
- steps_accel[ 9] = 384;
- steps_accel[10] = 512;
- steps_accel[11] = 768;
- steps_accel[12] = 1024;
+ qsort (steps_accel, STEPS_ACCEL_CNT, sizeof (u32), sort_by_u32);
+ qsort (steps_loops, STEPS_LOOPS_CNT, sizeof (u32), sort_by_u32);
// find out highest kernel-loops that stays below target_ms, we can use it later for multiplication as this is a linear function
u32 kernel_loops_tmp;
- for (kernel_loops_tmp = MIN (kernel_loops_max, 200); kernel_loops_tmp >= kernel_loops_min; kernel_loops_tmp >>= 1)
+ for (kernel_loops_tmp = kernel_loops_max; kernel_loops_tmp > kernel_loops_min; kernel_loops_tmp >>= 1)
{
const double exec_ms = try_run (device_param, kernel_accel_min, kernel_loops_tmp, 1);
if (exec_ms < target_ms) break;
-
- if (kernel_loops_tmp == kernel_loops_min) break;
}
// kernel-accel
- double e_best = 0;
-
- for (int i = 0; i < STEPS_ACCEL_CNT; i++)
+ if (kernel_accel_min < kernel_accel_max)
{
- const u32 kernel_accel_try = steps_accel[i];
+ double e_best = 0;
- if (kernel_accel_try < kernel_accel_min) continue;
- if (kernel_accel_try > kernel_accel_max) break;
+ for (int i = 0; i < STEPS_ACCEL_CNT; i++)
+ {
+ const u32 kernel_accel_try = steps_accel[i];
- const double exec_ms = try_run (device_param, kernel_accel_try, kernel_loops_tmp, 1);
+ if (kernel_accel_try < kernel_accel_min) continue;
+ if (kernel_accel_try > kernel_accel_max) break;
- if (exec_ms > target_ms) break;
+ const double exec_ms = try_run (device_param, kernel_accel_try, kernel_loops_tmp, 1);
- const double e = kernel_accel_try / exec_ms;
+ if (exec_ms > target_ms) break;
- if (e > e_best)
- {
- kernel_accel = kernel_accel_try;
+ const double e = kernel_accel_try / exec_ms;
- e_best = e;
+ if (e > e_best)
+ {
+ kernel_accel = kernel_accel_try;
+
+ e_best = e;
+ }
}
}
// kernel-loops final
- e_best = 0;
-
- for (int i = 0; i < STEPS_LOOPS_CNT; i++)
+ if (kernel_loops_min < kernel_loops_max)
{
- const u32 kernel_loops_try = steps_loops[i];
+ double e_best = 0;
- if (kernel_loops_try < kernel_loops_min) continue;
- if (kernel_loops_try > kernel_loops_max) break;
+ for (int i = 0; i < STEPS_LOOPS_CNT; i++)
+ {
+ const u32 kernel_loops_try = steps_loops[i];
- const double exec_ms = try_run (device_param, kernel_accel, kernel_loops_try, 1);
+ if (kernel_loops_try < kernel_loops_min) continue;
+ if (kernel_loops_try > kernel_loops_max) break;
- if (exec_ms > target_ms) break;
+ const double exec_ms = try_run (device_param, kernel_accel, kernel_loops_try, 1);
- const double e = kernel_loops_try / exec_ms;
+ if (exec_ms > target_ms) break;
- if (e > e_best)
- {
- kernel_loops = kernel_loops_try;
+ const double e = kernel_loops_try / exec_ms;
+
+ if (e > e_best)
+ {
+ kernel_loops = kernel_loops_try;
- e_best = e;
+ e_best = e;
+ }
}
}
// final balance
- const double exec_ms = try_run (device_param, kernel_accel, kernel_loops, 1);
+ u32 kernel_accel_best = kernel_accel;
+ u32 kernel_loops_best = kernel_loops;
+
+ u32 exec_best = -1;
- e_best = exec_ms;
+ if ((kernel_accel_min < kernel_accel_max) || (kernel_loops_min < kernel_loops_max))
+ {
+ const double exec_ms = try_run (device_param, kernel_accel_best, kernel_loops_best, 1);
+
+ exec_best = exec_ms;
+ }
- u32 kernel_accel_try = kernel_accel;
- u32 kernel_loops_try = kernel_loops;
+ // reset
- for (int i = 0; i < 2; i++)
+ if (kernel_accel_min < kernel_accel_max)
{
- kernel_accel_try >>= 1;
- kernel_loops_try <<= 1;
+ u32 kernel_accel_try = kernel_accel;
+ u32 kernel_loops_try = kernel_loops;
- if (kernel_accel_try < kernel_accel_min) break;
- if (kernel_loops_try > kernel_loops_max) break;
+ for (int i = 0; i < 2; i++)
+ {
+ kernel_accel_try >>= 1;
+ kernel_loops_try <<= 1;
- const double exec_ms = try_run (device_param, kernel_accel_try, kernel_loops_try, 1);
+ if (kernel_accel_try < kernel_accel_min) break;
+ if (kernel_loops_try > kernel_loops_max) break;
- if (exec_ms > e_best) break;
+ const double exec_ms = try_run (device_param, kernel_accel_try, kernel_loops_try, 1);
- kernel_accel = kernel_accel_try;
- kernel_loops = kernel_loops_try;
+ if (exec_ms < exec_best)
+ {
+ kernel_accel_best = kernel_accel_try;
+ kernel_loops_best = kernel_loops_try;
- e_best = exec_ms;
+ exec_best = exec_ms;
+ }
+ }
}
- kernel_accel_try = kernel_accel;
- kernel_loops_try = kernel_loops;
+ // reset
- for (int i = 0; i < 2; i++)
+ if (kernel_loops_min < kernel_loops_max)
{
- kernel_accel_try <<= 1;
- kernel_loops_try >>= 1;
+ u32 kernel_accel_try = kernel_accel;
+ u32 kernel_loops_try = kernel_loops;
- if (kernel_accel_try > kernel_accel_max) break;
- if (kernel_loops_try < kernel_loops_min) break;
+ for (int i = 0; i < 2; i++)
+ {
+ kernel_accel_try <<= 1;
+ kernel_loops_try >>= 1;
- const double exec_ms = try_run (device_param, kernel_accel_try, kernel_loops_try, 1);
+ if (kernel_accel_try > kernel_accel_max) break;
+ if (kernel_loops_try < kernel_loops_min) break;
- if (exec_ms > e_best) break;
+ const double exec_ms = try_run (device_param, kernel_accel_try, kernel_loops_try, 1);
- kernel_accel = kernel_accel_try;
- kernel_loops = kernel_loops_try;
+ if (exec_ms < exec_best)
+ {
+ kernel_accel_best = kernel_accel_try;
+ kernel_loops_best = kernel_loops_try;
- e_best = exec_ms;
+ exec_best = exec_ms;
+ }
+ }
}
// reset timer
memset (device_param->exec_ms, 0, EXEC_CACHE * sizeof (double));
- // reset fake words
-
- for (u32 i = 0; i < kernel_power_max; i++)
- {
- device_param->pws_buf[i].pw_len = 0;
- }
-
- hc_clEnqueueWriteBuffer (data.ocl, device_param->command_queue, device_param->d_pws_buf, CL_TRUE, 0, device_param->size_pws, device_param->pws_buf, 0, NULL, NULL);
- hc_clEnqueueWriteBuffer (data.ocl, device_param->command_queue, device_param->d_pws_amp_buf, CL_TRUE, 0, device_param->size_pws, device_param->pws_buf, 0, NULL, NULL);
-
// store
- device_param->kernel_loops = kernel_loops;
+ kernel_accel = kernel_accel_best;
+ kernel_loops = kernel_loops_best;
+
device_param->kernel_accel = kernel_accel;
+ device_param->kernel_loops = kernel_loops;
const u32 kernel_power = device_param->device_processors * device_param->kernel_threads * device_param->kernel_accel;
device_param->kernel_power = kernel_power;
- log_info ("Device #%u: autotuned kernel-accel to %u", device_param->device_id + 1, kernel_accel);
- log_info ("Device #%u: autotuned kernel-loops to %u", device_param->device_id + 1, kernel_loops);
- log_info ("");
+ #ifdef DEBUG
+
+ if (data.quiet == 0)
+ {
+ clear_prompt ();
+
+ log_info ("Device #%u: autotuned kernel-accel to %u\n"
+ "Device #%u: autotuned kernel-loops to %u\n",
+ device_param->device_id + 1,
+ kernel_accel,
+ device_param->device_id + 1,
+ kernel_loops);
+
+ fprintf (stdout, "%s", PROMPT);
+ fflush (stdout);
+ }
+
+ #endif
}
-static void run_cracker (hc_device_param_t *device_param, const uint pw_cnt, const uint pws_cnt)
+static void run_cracker (hc_device_param_t *device_param, const uint pws_cnt)
{
+ char *line_buf = (char *) mymalloc (HCBUFSIZ);
+
// init speed timer
uint speed_pos = device_param->speed_pos;
+ device_param->kernel_params_mp_l_buf32[5];
}
- // bitslice optimization stuff
-
- if (data.attack_mode == ATTACK_MODE_BF)
- {
- if (data.opts_type & OPTS_TYPE_PT_BITSLICE)
- {
- run_kernel_tb (device_param, pws_cnt);
- }
- }
-
// iteration type
uint innerloop_step = 0;
if (data.devices_status == STATUS_QUIT) break;
if (data.devices_status == STATUS_BYPASS) break;
- if (data.salts_shown[salt_pos] == 1) continue;
-
salt_t *salt_buf = &data.salts_buf[salt_pos];
device_param->kernel_params_buf32[24] = salt_pos;
continue;
}
+ if (data.salts_shown[salt_pos] == 1)
+ {
+ data.words_progress_done[salt_pos] += (u64) pws_cnt * (u64) innerloop_left;
+
+ continue;
+ }
+
// initialize amplifiers
if (data.attack_mode == ATTACK_MODE_COMBI)
{
- char line_buf[BUFSIZ] = { 0 };
-
uint i = 0;
while (i < innerloop_left)
if (rule_len_out < 0)
{
- data.words_progress_rejected[salt_pos] += pw_cnt;
+ data.words_progress_rejected[salt_pos] += pws_cnt;
continue;
}
hc_clEnqueueCopyBuffer (data.ocl, device_param->command_queue, device_param->d_combs, device_param->d_combs_c, 0, 0, innerloop_left * sizeof (comb_t), 0, NULL, NULL);
}
- if (data.attack_exec == ATTACK_EXEC_INSIDE_KERNEL)
- {
- if (data.attack_mode == ATTACK_MODE_BF)
- {
- if (data.opts_type & OPTS_TYPE_PT_BITSLICE)
- {
- const uint size_tm = 32 * sizeof (bs_word_t);
+ choose_kernel (device_param, data.attack_exec, data.attack_mode, data.opts_type, salt_buf, highest_pw_len, pws_cnt);
- run_kernel_bzero (device_param, device_param->d_tm_c, size_tm);
-
- run_kernel_tm (device_param);
-
- hc_clEnqueueCopyBuffer (data.ocl, device_param->command_queue, device_param->d_tm_c, device_param->d_bfs_c, 0, 0, size_tm, 0, NULL, NULL);
- }
- }
-
- if (highest_pw_len < 16)
- {
- run_kernel (KERN_RUN_1, device_param, pws_cnt, true);
- }
- else if (highest_pw_len < 32)
- {
- run_kernel (KERN_RUN_2, device_param, pws_cnt, true);
- }
- else
- {
- run_kernel (KERN_RUN_3, device_param, pws_cnt, true);
- }
- }
- else
+ if (data.benchmark == 1)
{
- run_kernel_amp (device_param, pws_cnt);
+ double exec_ms_avg_prev = get_avg_exec_time (device_param, EXEC_CACHE);
- run_kernel (KERN_RUN_1, device_param, pws_cnt, false);
+ // a few caching rounds
- if (data.opts_type & OPTS_TYPE_HOOK12)
+ for (u32 i = 0; i < 2; i++)
{
- run_kernel (KERN_RUN_12, device_param, pws_cnt, false);
- }
+ hc_timer_set (&device_param->timer_speed);
- uint iter = salt_buf->salt_iter;
+ choose_kernel (device_param, data.attack_exec, data.attack_mode, data.opts_type, salt_buf, highest_pw_len, pws_cnt);
- uint loop_step = device_param->kernel_loops;
-
- for (uint loop_pos = 0; loop_pos < iter; loop_pos += loop_step)
- {
- uint loop_left = iter - loop_pos;
+ double exec_ms_avg = get_avg_exec_time (device_param, EXEC_CACHE);
- loop_left = MIN (loop_left, loop_step);
-
- device_param->kernel_params_buf32[25] = loop_pos;
- device_param->kernel_params_buf32[26] = loop_left;
-
- run_kernel (KERN_RUN_2, device_param, pws_cnt, true);
-
- if (data.devices_status == STATUS_CRACKED) break;
- if (data.devices_status == STATUS_ABORTED) break;
- if (data.devices_status == STATUS_QUIT) break;
+ exec_ms_avg_prev = exec_ms_avg;
}
- if (data.opts_type & OPTS_TYPE_HOOK23)
+ // benchmark_repeats became a maximum possible repeats
+
+ for (u32 i = 2; i < data.benchmark_repeats; i++)
{
- run_kernel (KERN_RUN_23, device_param, pws_cnt, false);
+ hc_timer_set (&device_param->timer_speed);
- hc_clEnqueueReadBuffer (data.ocl, device_param->command_queue, device_param->d_hooks, CL_TRUE, 0, device_param->size_hooks, device_param->hooks_buf, 0, NULL, NULL);
+ choose_kernel (device_param, data.attack_exec, data.attack_mode, data.opts_type, salt_buf, highest_pw_len, pws_cnt);
- // do something with data
+ double exec_ms_avg = get_avg_exec_time (device_param, EXEC_CACHE);
- hc_clEnqueueWriteBuffer (data.ocl, device_param->command_queue, device_param->d_hooks, CL_TRUE, 0, device_param->size_hooks, device_param->hooks_buf, 0, NULL, NULL);
- }
+ if ((exec_ms_avg_prev / exec_ms_avg) < 1.001) break;
- run_kernel (KERN_RUN_3, device_param, pws_cnt, false);
+ exec_ms_avg_prev = exec_ms_avg;
+ }
}
if (data.devices_status == STATUS_STOP_AT_CHECKPOINT) check_checkpoint ();
* progress
*/
- u64 perf_sum_all = (u64) pw_cnt * (u64) innerloop_left;
+ u64 perf_sum_all = (u64) pws_cnt * (u64) innerloop_left;
hc_thread_mutex_lock (mux_counter);
* speed
*/
- float speed_ms;
+ double speed_ms;
hc_timer_get (device_param->timer_speed, speed_ms);
hc_thread_mutex_lock (mux_display);
+ // current speed
+
device_param->speed_cnt[speed_pos] = perf_sum_all;
device_param->speed_ms[speed_pos] = speed_ms;
- device_param->speed_rec[speed_pos] = device_param->timer_speed;
-
hc_thread_mutex_unlock (mux_display);
speed_pos++;
* benchmark
*/
- if (data.benchmark == 1)
- {
- data.devices_status = STATUS_BYPASS;
- };
+ if (data.benchmark == 1) break;
}
}
device_param->speed_pos = speed_pos;
+
+ myfree (line_buf);
}
static void load_segment (wl_data_t *wl_data, FILE *fd)
if (feof (fd))
{
- fprintf (stderr, "bug!!\n");
+ fprintf (stderr, "BUG feof()!!\n");
return;
}
if (data.quiet == 0) log_info_nn ("Generating dictionary stats for %s: %llu bytes (%.2f%%), %llu words, %llu keyspace", dictfile, (unsigned long long int) comp, percent * 100, (unsigned long long int) cnt2, (unsigned long long int) cnt);
- time (&prev);
- }
-
- if (data.quiet == 0) log_info ("Generated dictionary stats for %s: %llu bytes, %llu words, %llu keyspace", dictfile, (unsigned long long int) comp, (unsigned long long int) cnt2, (unsigned long long int) cnt);
- if (data.quiet == 0) log_info ("");
-
- lsearch (&d, dictstat_base, dictstat_nmemb, sizeof (dictstat_t), sort_by_dictstat);
-
- hc_signal (sigHandler_default);
-
- return (cnt);
-}
-
-static void pw_transpose_to_hi1 (const pw_t *p1, pw_t *p2)
-{
- memcpy (p2->h.hi1, p1->h.hi1, 64 * sizeof (uint));
-}
-
-static uint pw_add_to_hc1 (hc_device_param_t *device_param, const u8 *pw_buf, const uint pw_len)
-{
- if (data.devices_status == STATUS_BYPASS) return 0;
-
- pw_cache_t *pw_cache = device_param->pw_caches + pw_len;
-
- uint cache_cnt = pw_cache->cnt;
-
- u8 *pw_hc1 = pw_cache->pw_buf.h.hc1[cache_cnt];
-
- memcpy (pw_hc1, pw_buf, pw_len);
-
- memset (pw_hc1 + pw_len, 0, 256 - pw_len);
-
- uint pws_cnt = device_param->pws_cnt;
-
- cache_cnt++;
-
- pw_t *pw = device_param->pws_buf + pws_cnt;
-
- device_param->pw_transpose (&pw_cache->pw_buf, pw);
-
- pw->pw_len = pw_len;
-
- pws_cnt++;
+ time (&prev);
+ }
- device_param->pws_cnt = pws_cnt;
- device_param->pw_cnt = pws_cnt * 1;
+ if (data.quiet == 0) log_info ("Generated dictionary stats for %s: %llu bytes, %llu words, %llu keyspace", dictfile, (unsigned long long int) comp, (unsigned long long int) cnt2, (unsigned long long int) cnt);
+ if (data.quiet == 0) log_info ("");
- cache_cnt = 0;
+ lsearch (&d, dictstat_base, dictstat_nmemb, sizeof (dictstat_t), sort_by_dictstat);
- pw_cache->cnt = cache_cnt;
+ hc_signal (sigHandler_default);
- return pws_cnt;
+ return (cnt);
}
static void *thread_monitor (void *p)
fseek (fp, out_info[j].seek, SEEK_SET);
+ char *line_buf = (char *) mymalloc (HCBUFSIZ);
+
while (!feof (fp))
{
- char line_buf[BUFSIZ] = { 0 };
-
- char *ptr = fgets (line_buf, BUFSIZ - 1, fp);
+ char *ptr = fgets (line_buf, HCBUFSIZ - 1, fp);
if (ptr == NULL) break;
wpa_t *wpas = (wpa_t *) data.esalts_buf;
wpa_t *wpa = &wpas[salt_pos];
- uint pke[25] = { 0 };
-
- char *pke_ptr = (char *) pke;
-
- for (uint i = 0; i < 25; i++)
- {
- pke[i] = byte_swap_32 (wpa->pke[i]);
- }
-
- u8 mac1[6] = { 0 };
- u8 mac2[6] = { 0 };
-
- memcpy (mac1, pke_ptr + 23, 6);
- memcpy (mac2, pke_ptr + 29, 6);
-
// compare hex string(s) vs binary MAC address(es)
for (uint i = 0, j = 0; i < 6; i++, j += 2)
{
- if (mac1[i] != hex_to_u8 ((const u8 *) &mac1_pos[j]))
+ if (wpa->orig_mac1[i] != hex_to_u8 ((const u8 *) &mac1_pos[j]))
{
cracked = 0;
+
break;
}
}
for (uint i = 0, j = 0; i < 6; i++, j += 2)
{
- if (mac2[i] != hex_to_u8 ((const u8 *) &mac2_pos[j]))
+ if (wpa->orig_mac2[i] != hex_to_u8 ((const u8 *) &mac2_pos[j]))
{
cracked = 0;
+
break;
}
}
if (data.devices_status == STATUS_CRACKED) break;
}
+ myfree (line_buf);
+
out_info[j].seek = ftell (fp);
//hc_thread_mutex_unlock (mux_display);
return (p);
}
-static uint get_work (hc_device_param_t *device_param, const u64 max)
+static void pw_add (hc_device_param_t *device_param, const u8 *pw_buf, const int pw_len)
+{
+ if (device_param->pws_cnt < device_param->kernel_power)
+ {
+ pw_t *pw = (pw_t *) device_param->pws_buf + device_param->pws_cnt;
+
+ u8 *ptr = (u8 *) pw->i;
+
+ memcpy (ptr, pw_buf, pw_len);
+
+ memset (ptr + pw_len, 0, sizeof (pw->i) - pw_len);
+
+ pw->pw_len = pw_len;
+
+ device_param->pws_cnt++;
+ }
+ else
+ {
+ fprintf (stderr, "BUG pw_add()!!\n");
+
+ return;
+ }
+}
+
+static uint get_work (hc_device_param_t *device_param, const u64 max, const bool allow_div)
{
hc_thread_mutex_lock (mux_dispatcher);
const u64 words_left = words_base - words_cur;
- if (data.kernel_power_all > words_left)
+ if (allow_div)
{
- if (data.kernel_power_div == 0)
+ if (data.kernel_power_all > words_left)
{
- data.kernel_power_div = find_kernel_power_div (words_left, data.kernel_power_all);
+ if (data.kernel_power_div == 0)
+ {
+ data.kernel_power_div = find_kernel_power_div (words_left, data.kernel_power_all);
+ }
}
- }
- if (data.kernel_power_div)
- {
- if (device_param->kernel_power == device_param->kernel_power_user)
+ if (data.kernel_power_div)
{
- const u32 kernel_power_new = (float) device_param->kernel_power * data.kernel_power_div;
-
- if (kernel_power_new < device_param->kernel_power)
+ if (device_param->kernel_power == device_param->kernel_power_user)
{
- device_param->kernel_power = kernel_power_new;
+ const u32 kernel_power_new = (float) device_param->kernel_power * data.kernel_power_div;
+
+ if (kernel_power_new < device_param->kernel_power)
+ {
+ device_param->kernel_power = kernel_power_new;
+ }
}
}
}
if (device_param->skipped) return NULL;
- if ((device_param->kernel_accel == 0) && (device_param->kernel_loops == 0))
- {
- autotune (device_param);
- }
+ autotune (device_param);
+
+ char *buf = (char *) mymalloc (HCBUFSIZ);
const uint attack_kern = data.attack_kern;
while (words_cur < kernel_power)
{
- char buf[BUFSIZ] = { 0 };
-
- char *line_buf = fgets (buf, sizeof (buf), stdin);
+ char *line_buf = fgets (buf, HCBUFSIZ - 1, stdin);
if (line_buf == NULL) break;
}
}
- device_param->pw_add (device_param, (u8 *) line_buf, line_len);
+ pw_add (device_param, (u8 *) line_buf, line_len);
words_cur++;
if (data.devices_status == STATUS_QUIT) break;
if (data.devices_status == STATUS_BYPASS) break;
- // we need 2 flushing because we have two independant caches and it can occur
- // that one buffer is already at threshold plus for that length also exists
- // more data in the 2nd buffer so it would overflow
+ // flush
- // flush session 1
+ const uint pws_cnt = device_param->pws_cnt;
+ if (pws_cnt)
{
- for (int pw_len = 0; pw_len < PW_MAX1; pw_len++)
- {
- pw_cache_t *pw_cache = &device_param->pw_caches[pw_len];
-
- const uint pw_cache_cnt = pw_cache->cnt;
-
- if (pw_cache_cnt == 0) continue;
-
- pw_cache->cnt = 0;
-
- uint pws_cnt = device_param->pws_cnt;
-
- pw_t *pw = device_param->pws_buf + pws_cnt;
-
- device_param->pw_transpose (&pw_cache->pw_buf, pw);
-
- pw->pw_len = pw_len;
-
- uint pw_cnt = device_param->pw_cnt;
-
- pw_cnt += pw_cache_cnt;
-
- device_param->pw_cnt = pw_cnt;
-
- pws_cnt++;
-
- device_param->pws_cnt = pws_cnt;
-
- if (pws_cnt == device_param->kernel_power_user) break;
- }
-
- const uint pw_cnt = device_param->pw_cnt;
- const uint pws_cnt = device_param->pws_cnt;
-
- if (pws_cnt)
- {
- run_copy (device_param, pws_cnt);
-
- run_cracker (device_param, pw_cnt, pws_cnt);
+ run_copy (device_param, pws_cnt);
- device_param->pw_cnt = 0;
- device_param->pws_cnt = 0;
- }
- }
+ run_cracker (device_param, pws_cnt);
- // flush session 2
+ device_param->pws_cnt = 0;
- {
- for (int pw_len = 0; pw_len < PW_MAX1; pw_len++)
+ if (attack_kern == ATTACK_KERN_STRAIGHT)
{
- pw_cache_t *pw_cache = &device_param->pw_caches[pw_len];
-
- const uint pw_cache_cnt = pw_cache->cnt;
-
- if (pw_cache_cnt == 0) continue;
-
- pw_cache->cnt = 0;
-
- uint pws_cnt = device_param->pws_cnt;
-
- pw_t *pw = device_param->pws_buf + pws_cnt;
-
- device_param->pw_transpose (&pw_cache->pw_buf, pw);
-
- pw->pw_len = pw_len;
-
- uint pw_cnt = device_param->pw_cnt;
-
- pw_cnt += pw_cache_cnt;
-
- device_param->pw_cnt = pw_cnt;
-
- pws_cnt++;
-
- device_param->pws_cnt = pws_cnt;
+ run_kernel_bzero (device_param, device_param->d_rules_c, device_param->size_rules_c);
}
-
- const uint pw_cnt = device_param->pw_cnt;
- const uint pws_cnt = device_param->pws_cnt;
-
- if (pws_cnt)
+ else if (attack_kern == ATTACK_KERN_COMBI)
{
- run_copy (device_param, pws_cnt);
-
- run_cracker (device_param, pw_cnt, pws_cnt);
-
- device_param->pw_cnt = 0;
- device_param->pws_cnt = 0;
+ run_kernel_bzero (device_param, device_param->d_combs_c, device_param->size_combs);
}
}
}
+ device_param->kernel_accel = 0;
+ device_param->kernel_loops = 0;
+
+ myfree (buf);
+
return NULL;
}
if (device_param->skipped) return NULL;
- if ((device_param->kernel_accel == 0) && (device_param->kernel_loops == 0))
- {
- autotune (device_param);
- }
+ autotune (device_param);
const uint attack_mode = data.attack_mode;
const uint attack_kern = data.attack_kern;
{
while ((data.devices_status != STATUS_EXHAUSTED) && (data.devices_status != STATUS_CRACKED) && (data.devices_status != STATUS_ABORTED) && (data.devices_status != STATUS_QUIT))
{
- const uint work = get_work (device_param, -1);
+ const uint work = get_work (device_param, -1, true);
if (work == 0) break;
const u64 words_off = device_param->words_off;
const u64 words_fin = words_off + work;
- const uint pw_cnt = work;
const uint pws_cnt = work;
- device_param->pw_cnt = pw_cnt;
device_param->pws_cnt = pws_cnt;
if (pws_cnt)
{
run_copy (device_param, pws_cnt);
- run_cracker (device_param, pw_cnt, pws_cnt);
+ run_cracker (device_param, pws_cnt);
- device_param->pw_cnt = 0;
device_param->pws_cnt = 0;
+
+ run_kernel_bzero (device_param, device_param->d_bfs_c, device_param->size_bfs);
}
if (data.devices_status == STATUS_STOP_AT_CHECKPOINT) check_checkpoint ();
if (data.devices_status == STATUS_QUIT) break;
if (data.devices_status == STATUS_BYPASS) break;
+ if (data.benchmark == 1) break;
+
device_param->words_done = words_fin;
}
}
u64 words_off = 0;
u64 words_fin = 0;
+ bool allow_div = true;
+
u64 max = -1;
while (max)
{
- const uint work = get_work (device_param, max);
+ const uint work = get_work (device_param, max, allow_div);
+
+ allow_div = false;
if (work == 0) break;
}
}
- device_param->pw_add (device_param, (u8 *) line_buf, line_len);
+ pw_add (device_param, (u8 *) line_buf, line_len);
if (data.devices_status == STATUS_STOP_AT_CHECKPOINT) check_checkpoint ();
if (data.devices_status == STATUS_QUIT) break;
if (data.devices_status == STATUS_BYPASS) break;
- // we need 2 flushing because we have two independant caches and it can occur
- // that one buffer is already at threshold plus for that length also exists
- // more data in the 2nd buffer so it would overflow
-
//
- // flush session 1
+ // flush
//
- {
- for (int pw_len = 0; pw_len < PW_MAX1; pw_len++)
- {
- pw_cache_t *pw_cache = &device_param->pw_caches[pw_len];
-
- const uint pw_cache_cnt = pw_cache->cnt;
-
- if (pw_cache_cnt == 0) continue;
-
- pw_cache->cnt = 0;
-
- uint pws_cnt = device_param->pws_cnt;
-
- pw_t *pw = device_param->pws_buf + pws_cnt;
-
- device_param->pw_transpose (&pw_cache->pw_buf, pw);
-
- pw->pw_len = pw_len;
-
- uint pw_cnt = device_param->pw_cnt;
-
- pw_cnt += pw_cache_cnt;
-
- device_param->pw_cnt = pw_cnt;
-
- pws_cnt++;
-
- device_param->pws_cnt = pws_cnt;
-
- if (pws_cnt == device_param->kernel_power_user) break;
- }
-
- const uint pw_cnt = device_param->pw_cnt;
- const uint pws_cnt = device_param->pws_cnt;
-
- if (pws_cnt)
- {
- run_copy (device_param, pws_cnt);
-
- run_cracker (device_param, pw_cnt, pws_cnt);
-
- device_param->pw_cnt = 0;
- device_param->pws_cnt = 0;
- }
-
- if (data.devices_status == STATUS_STOP_AT_CHECKPOINT) check_checkpoint ();
-
- if (data.devices_status == STATUS_CRACKED) break;
- if (data.devices_status == STATUS_ABORTED) break;
- if (data.devices_status == STATUS_QUIT) break;
- if (data.devices_status == STATUS_BYPASS) break;
- }
-
- //
- // flush session 2
- //
+ const uint pws_cnt = device_param->pws_cnt;
+ if (pws_cnt)
{
- for (int pw_len = 0; pw_len < PW_MAX1; pw_len++)
- {
- pw_cache_t *pw_cache = &device_param->pw_caches[pw_len];
-
- const uint pw_cache_cnt = pw_cache->cnt;
-
- if (pw_cache_cnt == 0) continue;
-
- pw_cache->cnt = 0;
-
- uint pws_cnt = device_param->pws_cnt;
-
- pw_t *pw = device_param->pws_buf + pws_cnt;
-
- device_param->pw_transpose (&pw_cache->pw_buf, pw);
-
- pw->pw_len = pw_len;
-
- uint pw_cnt = device_param->pw_cnt;
-
- pw_cnt += pw_cache_cnt;
+ run_copy (device_param, pws_cnt);
- device_param->pw_cnt = pw_cnt;
+ run_cracker (device_param, pws_cnt);
- pws_cnt++;
+ device_param->pws_cnt = 0;
- device_param->pws_cnt = pws_cnt;
+ if (attack_kern == ATTACK_KERN_STRAIGHT)
+ {
+ run_kernel_bzero (device_param, device_param->d_rules_c, device_param->size_rules_c);
}
-
- const uint pw_cnt = device_param->pw_cnt;
- const uint pws_cnt = device_param->pws_cnt;
-
- if (pws_cnt)
+ else if (attack_kern == ATTACK_KERN_COMBI)
{
- run_copy (device_param, pws_cnt);
-
- run_cracker (device_param, pw_cnt, pws_cnt);
-
- device_param->pw_cnt = 0;
- device_param->pws_cnt = 0;
+ run_kernel_bzero (device_param, device_param->d_combs_c, device_param->size_combs);
}
+ }
- if (data.devices_status == STATUS_STOP_AT_CHECKPOINT) check_checkpoint ();
+ if (data.devices_status == STATUS_STOP_AT_CHECKPOINT) check_checkpoint ();
- if (data.devices_status == STATUS_CRACKED) break;
- if (data.devices_status == STATUS_ABORTED) break;
- if (data.devices_status == STATUS_QUIT) break;
- if (data.devices_status == STATUS_BYPASS) break;
- }
+ if (data.devices_status == STATUS_CRACKED) break;
+ if (data.devices_status == STATUS_ABORTED) break;
+ if (data.devices_status == STATUS_QUIT) break;
+ if (data.devices_status == STATUS_BYPASS) break;
if (words_fin == 0) break;
fclose (fd);
}
+ device_param->kernel_accel = 0;
+ device_param->kernel_loops = 0;
+
return NULL;
}
// hlfmt hashcat
-static void hlfmt_hash_hashcat (char line_buf[BUFSIZ], int line_len, char **hashbuf_pos, int *hashbuf_len)
+static void hlfmt_hash_hashcat (char *line_buf, int line_len, char **hashbuf_pos, int *hashbuf_len)
{
if (data.username == 0)
{
}
}
-static void hlfmt_user_hashcat (char line_buf[BUFSIZ], int line_len, char **userbuf_pos, int *userbuf_len)
+static void hlfmt_user_hashcat (char *line_buf, int line_len, char **userbuf_pos, int *userbuf_len)
{
char *pos = NULL;
int len = 0;
// hlfmt pwdump
-static int hlfmt_detect_pwdump (char line_buf[BUFSIZ], int line_len)
+static int hlfmt_detect_pwdump (char *line_buf, int line_len)
{
int sep_cnt = 0;
return 0;
}
-static void hlfmt_hash_pwdump (char line_buf[BUFSIZ], int line_len, char **hashbuf_pos, int *hashbuf_len)
+static void hlfmt_hash_pwdump (char *line_buf, int line_len, char **hashbuf_pos, int *hashbuf_len)
{
char *pos = NULL;
int len = 0;
*hashbuf_len = len;
}
-static void hlfmt_user_pwdump (char line_buf[BUFSIZ], int line_len, char **userbuf_pos, int *userbuf_len)
+static void hlfmt_user_pwdump (char *line_buf, int line_len, char **userbuf_pos, int *userbuf_len)
{
char *pos = NULL;
int len = 0;
// hlfmt passwd
-static int hlfmt_detect_passwd (char line_buf[BUFSIZ], int line_len)
+static int hlfmt_detect_passwd (char *line_buf, int line_len)
{
int sep_cnt = 0;
return 0;
}
-static void hlfmt_hash_passwd (char line_buf[BUFSIZ], int line_len, char **hashbuf_pos, int *hashbuf_len)
+static void hlfmt_hash_passwd (char *line_buf, int line_len, char **hashbuf_pos, int *hashbuf_len)
{
char *pos = NULL;
int len = 0;
*hashbuf_len = len;
}
-static void hlfmt_user_passwd (char line_buf[BUFSIZ], int line_len, char **userbuf_pos, int *userbuf_len)
+static void hlfmt_user_passwd (char *line_buf, int line_len, char **userbuf_pos, int *userbuf_len)
{
char *pos = NULL;
int len = 0;
// hlfmt shadow
-static int hlfmt_detect_shadow (char line_buf[BUFSIZ], int line_len)
+static int hlfmt_detect_shadow (char *line_buf, int line_len)
{
int sep_cnt = 0;
return 0;
}
-static void hlfmt_hash_shadow (char line_buf[BUFSIZ], int line_len, char **hashbuf_pos, int *hashbuf_len)
+static void hlfmt_hash_shadow (char *line_buf, int line_len, char **hashbuf_pos, int *hashbuf_len)
{
hlfmt_hash_passwd (line_buf, line_len, hashbuf_pos, hashbuf_len);
}
-static void hlfmt_user_shadow (char line_buf[BUFSIZ], int line_len, char **userbuf_pos, int *userbuf_len)
+static void hlfmt_user_shadow (char *line_buf, int line_len, char **userbuf_pos, int *userbuf_len)
{
hlfmt_user_passwd (line_buf, line_len, userbuf_pos, userbuf_len);
}
// hlfmt main
-static void hlfmt_hash (uint hashfile_format, char line_buf[BUFSIZ], int line_len, char **hashbuf_pos, int *hashbuf_len)
+static void hlfmt_hash (uint hashfile_format, char *line_buf, int line_len, char **hashbuf_pos, int *hashbuf_len)
{
switch (hashfile_format)
{
}
}
-static void hlfmt_user (uint hashfile_format, char line_buf[BUFSIZ], int line_len, char **userbuf_pos, int *userbuf_len)
+static void hlfmt_user (uint hashfile_format, char *line_buf, int line_len, char **userbuf_pos, int *userbuf_len)
{
switch (hashfile_format)
{
}
}
+char *strhlfmt (const uint hashfile_format)
+{
+ switch (hashfile_format)
+ {
+ case HLFMT_HASHCAT: return ((char *) HLFMT_TEXT_HASHCAT); break;
+ case HLFMT_PWDUMP: return ((char *) HLFMT_TEXT_PWDUMP); break;
+ case HLFMT_PASSWD: return ((char *) HLFMT_TEXT_PASSWD); break;
+ case HLFMT_SHADOW: return ((char *) HLFMT_TEXT_SHADOW); break;
+ case HLFMT_DCC: return ((char *) HLFMT_TEXT_DCC); break;
+ case HLFMT_DCC2: return ((char *) HLFMT_TEXT_DCC2); break;
+ case HLFMT_NETNTLM1: return ((char *) HLFMT_TEXT_NETNTLM1); break;
+ case HLFMT_NETNTLM2: return ((char *) HLFMT_TEXT_NETNTLM2); break;
+ case HLFMT_NSLDAP: return ((char *) HLFMT_TEXT_NSLDAP); break;
+ case HLFMT_NSLDAPS: return ((char *) HLFMT_TEXT_NSLDAPS); break;
+ }
+
+ return ((char *) "Unknown");
+}
+
static uint hlfmt_detect (FILE *fp, uint max_check)
{
// Exception: those formats are wrongly detected as HLFMT_SHADOW, prevent it
uint num_check = 0;
+ char *line_buf = (char *) mymalloc (HCBUFSIZ);
+
while (!feof (fp))
{
- char line_buf[BUFSIZ] = { 0 };
-
int line_len = fgetl (fp, line_buf);
if (line_len == 0) continue;
num_check++;
}
+ myfree (line_buf);
+
uint hashlist_format = HLFMT_HASHCAT;
for (int i = 1; i < HLFMTS_CNT; i++)
if (getenv ("GPU_USE_SYNC_OBJECTS") == NULL)
putenv ((char *) "GPU_USE_SYNC_OBJECTS=1");
+ if (getenv ("CUDA_CACHE_DISABLE") == NULL)
+ putenv ((char *) "CUDA_CACHE_DISABLE=1");
+
/**
* Real init
*/
uint version = VERSION;
uint quiet = QUIET;
uint benchmark = BENCHMARK;
+ uint benchmark_repeats = BENCHMARK_REPEATS;
uint show = SHOW;
uint left = LEFT;
uint username = USERNAME;
u64 limit = LIMIT;
uint keyspace = KEYSPACE;
uint potfile_disable = POTFILE_DISABLE;
+ char *potfile_path = NULL;
uint debug_mode = DEBUG_MODE;
char *debug_file = NULL;
char *induction_dir = NULL;
#define IDX_LIMIT 'l'
#define IDX_KEYSPACE 0xff35
#define IDX_POTFILE_DISABLE 0xff06
+ #define IDX_POTFILE_PATH 0xffe0
#define IDX_DEBUG_MODE 0xff43
#define IDX_DEBUG_FILE 0xff44
#define IDX_INDUCTION_DIR 0xff46
#define IDX_FORCE 0xff08
#define IDX_RUNTIME 0xff09
#define IDX_BENCHMARK 'b'
+ #define IDX_BENCHMARK_REPEATS 0xff78
#define IDX_HASH_MODE 'm'
#define IDX_ATTACK_MODE 'a'
#define IDX_RP_FILE 'r'
{"limit", required_argument, 0, IDX_LIMIT},
{"keyspace", no_argument, 0, IDX_KEYSPACE},
{"potfile-disable", no_argument, 0, IDX_POTFILE_DISABLE},
+ {"potfile-path", required_argument, 0, IDX_POTFILE_PATH},
{"debug-mode", required_argument, 0, IDX_DEBUG_MODE},
{"debug-file", required_argument, 0, IDX_DEBUG_FILE},
{"induction-dir", required_argument, 0, IDX_INDUCTION_DIR},
{"outfile-check-dir", required_argument, 0, IDX_OUTFILE_CHECK_DIR},
{"force", no_argument, 0, IDX_FORCE},
{"benchmark", no_argument, 0, IDX_BENCHMARK},
+ {"benchmark-repeats", required_argument, 0, IDX_BENCHMARK_REPEATS},
{"restore", no_argument, 0, IDX_RESTORE},
{"restore-disable", no_argument, 0, IDX_RESTORE_DISABLE},
{"status", no_argument, 0, IDX_STATUS},
case IDX_REMOVE_TIMER: remove_timer = atoi (optarg);
remove_timer_chgd = 1; break;
case IDX_POTFILE_DISABLE: potfile_disable = 1; break;
+ case IDX_POTFILE_PATH: potfile_path = optarg; break;
case IDX_DEBUG_MODE: debug_mode = atoi (optarg); break;
case IDX_DEBUG_FILE: debug_file = optarg; break;
case IDX_INDUCTION_DIR: induction_dir = optarg; break;
case IDX_LIMIT: limit = atoll (optarg); break;
case IDX_KEYSPACE: keyspace = 1; break;
case IDX_BENCHMARK: benchmark = 1; break;
+ case IDX_BENCHMARK_REPEATS: benchmark_repeats = atoi (optarg); break;
case IDX_RESTORE: break;
case IDX_RESTORE_DISABLE: restore_disable = 1; break;
case IDX_STATUS: status = 1; break;
return (-1);
}
- if (hash_mode_chgd && hash_mode > 13000) // just added to remove compiler warnings for hash_mode_chgd
+ if (hash_mode_chgd && hash_mode > 13400) // just added to remove compiler warnings for hash_mode_chgd
{
log_error ("ERROR: Invalid hash-type specified");
return (-1);
}
- if (kernel_accel_chgd == 1 && kernel_loops_chgd == 0)
- {
- log_error ("ERROR: If kernel-accel is specified, kernel-loops need to be specified as well");
-
- return (-1);
- }
-
- if (kernel_loops_chgd == 1 && kernel_accel_chgd == 0)
- {
- log_error ("ERROR: If kernel-loops is specified, kernel-accel need to be specified as well");
-
- return (-1);
- }
-
if (kernel_accel_chgd == 1)
{
if (kernel_accel < 1)
return (-1);
}
- if ((opencl_vector_width != 0) && (opencl_vector_width != 1) && (opencl_vector_width != 2) && (opencl_vector_width != 4) && (opencl_vector_width != 8))
+ if (opencl_vector_width_chgd && (!is_power_of_2(opencl_vector_width) || opencl_vector_width > 16))
{
log_error ("ERROR: opencl-vector-width %i not allowed", opencl_vector_width);
data.rp_gen_seed = rp_gen_seed;
data.force = force;
data.benchmark = benchmark;
+ data.benchmark_repeats = benchmark_repeats;
data.skip = skip;
data.limit = limit;
#if defined(HAVE_HWMON) && defined(HAVE_ADL)
logfile_top_uint (attack_mode);
logfile_top_uint (attack_kern);
logfile_top_uint (benchmark);
+ logfile_top_uint (benchmark_repeats);
logfile_top_uint (bitmap_min);
logfile_top_uint (bitmap_max);
logfile_top_uint (debug_mode);
logfile_top_uint (outfile_check_timer);
logfile_top_uint (outfile_format);
logfile_top_uint (potfile_disable);
+ logfile_top_string (potfile_path);
#if defined(HAVE_HWMON) && defined(HAVE_ADL)
logfile_top_uint (powertune_enable);
#endif
dgst_pos3 = 1;
break;
+ case 125: hash_type = HASH_TYPE_SHA1;
+ salt_type = SALT_TYPE_EMBEDDED;
+ attack_exec = ATTACK_EXEC_INSIDE_KERNEL;
+ opts_type = OPTS_TYPE_PT_GENERATE_BE
+ | OPTS_TYPE_PT_ADD80
+ | OPTS_TYPE_PT_ADDBITS15
+ | OPTS_TYPE_ST_HEX;
+ kern_type = KERN_TYPE_SHA1_SLTPW;
+ dgst_size = DGST_SIZE_4_5;
+ parse_func = arubaos_parse_hash;
+ sort_by_digest = sort_by_digest_4_5;
+ opti_type = OPTI_TYPE_ZERO_BYTE
+ | OPTI_TYPE_PRECOMPUTE_INIT
+ | OPTI_TYPE_PRECOMPUTE_MERKLE
+ | OPTI_TYPE_EARLY_SKIP
+ | OPTI_TYPE_NOT_ITERATED
+ | OPTI_TYPE_PREPENDED_SALT
+ | OPTI_TYPE_RAW_HASH;
+ dgst_pos0 = 3;
+ dgst_pos1 = 4;
+ dgst_pos2 = 2;
+ dgst_pos3 = 1;
+ break;
+
case 130: hash_type = HASH_TYPE_SHA1;
salt_type = SALT_TYPE_INTERN;
attack_exec = ATTACK_EXEC_INSIDE_KERNEL;
case 8300: hash_type = HASH_TYPE_SHA1;
salt_type = SALT_TYPE_EMBEDDED;
attack_exec = ATTACK_EXEC_INSIDE_KERNEL;
- opts_type = OPTS_TYPE_PT_GENERATE_LE
+ opts_type = OPTS_TYPE_PT_GENERATE_BE
| OPTS_TYPE_ST_HEX
| OPTS_TYPE_ST_ADD80;
kern_type = KERN_TYPE_NSEC3;
dgst_pos3 = 3;
break;
- case 12900: hash_type = HASH_TYPE_PBKDF2_SHA256;
+ case 12900: hash_type = HASH_TYPE_PBKDF2_SHA256;
+ salt_type = SALT_TYPE_EMBEDDED;
+ attack_exec = ATTACK_EXEC_OUTSIDE_KERNEL;
+ opts_type = OPTS_TYPE_PT_GENERATE_LE;
+ kern_type = KERN_TYPE_ANDROIDFDE_SAMSUNG;
+ dgst_size = DGST_SIZE_4_8;
+ parse_func = androidfde_samsung_parse_hash;
+ sort_by_digest = sort_by_digest_4_8;
+ opti_type = OPTI_TYPE_ZERO_BYTE;
+ dgst_pos0 = 0;
+ dgst_pos1 = 1;
+ dgst_pos2 = 2;
+ dgst_pos3 = 3;
+ break;
+
+ case 13000: hash_type = HASH_TYPE_PBKDF2_SHA256;
+ salt_type = SALT_TYPE_EMBEDDED;
+ attack_exec = ATTACK_EXEC_OUTSIDE_KERNEL;
+ opts_type = OPTS_TYPE_PT_GENERATE_LE;
+ kern_type = KERN_TYPE_RAR5;
+ dgst_size = DGST_SIZE_4_4;
+ parse_func = rar5_parse_hash;
+ sort_by_digest = sort_by_digest_4_4;
+ opti_type = OPTI_TYPE_ZERO_BYTE;
+ dgst_pos0 = 0;
+ dgst_pos1 = 1;
+ dgst_pos2 = 2;
+ dgst_pos3 = 3;
+ break;
+
+ case 13100: hash_type = HASH_TYPE_KRB5TGS;
+ salt_type = SALT_TYPE_EMBEDDED;
+ attack_exec = ATTACK_EXEC_INSIDE_KERNEL;
+ opts_type = OPTS_TYPE_PT_GENERATE_LE;
+ kern_type = KERN_TYPE_KRB5TGS;
+ dgst_size = DGST_SIZE_4_4;
+ parse_func = krb5tgs_parse_hash;
+ sort_by_digest = sort_by_digest_4_4;
+ opti_type = OPTI_TYPE_ZERO_BYTE
+ | OPTI_TYPE_NOT_ITERATED;
+ dgst_pos0 = 0;
+ dgst_pos1 = 1;
+ dgst_pos2 = 2;
+ dgst_pos3 = 3;
+ break;
+
+ case 13200: hash_type = HASH_TYPE_AES;
salt_type = SALT_TYPE_EMBEDDED;
attack_exec = ATTACK_EXEC_OUTSIDE_KERNEL;
opts_type = OPTS_TYPE_PT_GENERATE_LE;
- kern_type = KERN_TYPE_ANDROIDFDE_SAMSUNG;
- dgst_size = DGST_SIZE_4_8;
- parse_func = androidfde_samsung_parse_hash;
- sort_by_digest = sort_by_digest_4_8;
+ kern_type = KERN_TYPE_AXCRYPT;
+ dgst_size = DGST_SIZE_4_4;
+ parse_func = axcrypt_parse_hash;
+ sort_by_digest = sort_by_digest_4_4;
opti_type = OPTI_TYPE_ZERO_BYTE;
dgst_pos0 = 0;
dgst_pos1 = 1;
dgst_pos3 = 3;
break;
- case 13000: hash_type = HASH_TYPE_PBKDF2_SHA256;
+ case 13300: hash_type = HASH_TYPE_SHA1;
+ salt_type = SALT_TYPE_NONE;
+ attack_exec = ATTACK_EXEC_INSIDE_KERNEL;
+ opts_type = OPTS_TYPE_PT_GENERATE_BE
+ | OPTS_TYPE_PT_ADD80
+ | OPTS_TYPE_PT_ADDBITS15;
+ kern_type = KERN_TYPE_SHA1_AXCRYPT;
+ dgst_size = DGST_SIZE_4_5;
+ parse_func = sha1axcrypt_parse_hash;
+ sort_by_digest = sort_by_digest_4_5;
+ opti_type = OPTI_TYPE_ZERO_BYTE
+ | OPTI_TYPE_PRECOMPUTE_INIT
+ | OPTI_TYPE_EARLY_SKIP
+ | OPTI_TYPE_NOT_ITERATED
+ | OPTI_TYPE_NOT_SALTED;
+ dgst_pos0 = 0;
+ dgst_pos1 = 4;
+ dgst_pos2 = 3;
+ dgst_pos3 = 2;
+ break;
+
+ case 13400: hash_type = HASH_TYPE_AES;
salt_type = SALT_TYPE_EMBEDDED;
attack_exec = ATTACK_EXEC_OUTSIDE_KERNEL;
opts_type = OPTS_TYPE_PT_GENERATE_LE;
- kern_type = KERN_TYPE_RAR5;
+ kern_type = KERN_TYPE_KEEPASS;
dgst_size = DGST_SIZE_4_4;
- parse_func = rar5_parse_hash;
+ parse_func = keepass_parse_hash;
sort_by_digest = sort_by_digest_4_4;
opti_type = OPTI_TYPE_ZERO_BYTE;
dgst_pos0 = 0;
}
/**
- * transpose
+ * parser
*/
data.parse_func = parse_func;
case 12000: esalt_size = sizeof (pbkdf2_sha1_t); break;
case 12100: esalt_size = sizeof (pbkdf2_sha512_t); break;
case 13000: esalt_size = sizeof (rar5_t); break;
+ case 13100: esalt_size = sizeof (krb5tgs_t); break;
+ case 13400: esalt_size = sizeof (keepass_t); break;
}
data.esalt_size = esalt_size;
if (keyspace == 0)
{
- snprintf (dictstat, sizeof (dictstat) - 1, "%s/hashcat.dictstat", profile_dir);
+ snprintf (dictstat, sizeof (dictstat) - 1, "%s/%s", profile_dir, DICTSTAT_FILENAME);
dictstat_fp = fopen (dictstat, "rb");
char potfile[256] = { 0 };
- snprintf (potfile, sizeof (potfile) - 1, "%s/%s.pot", session_dir, session);
+ if (potfile_path == NULL)
+ {
+ snprintf (potfile, sizeof (potfile) - 1, "%s/%s", profile_dir, POTFILE_FILENAME);
+ }
+ else
+ {
+ strncpy (potfile, potfile_path, sizeof (potfile) - 1);
+ }
data.pot_fp = NULL;
uint line_num = 0;
+ char *line_buf = (char *) mymalloc (HCBUFSIZ);
+
while (!feof (pot_fp))
{
line_num++;
- char line_buf[BUFSIZ] = { 0 };
-
int line_len = fgetl (pot_fp, line_buf);
if (line_len == 0) continue;
pot_cnt++;
}
+ myfree (line_buf);
+
fclose (pot_fp);
SUPPRESS_OUTPUT = 0;
switch (hash_mode)
{
+ case 125: if (pw_max > 32) pw_max = 32;
+ break;
case 400: if (pw_max > 40) pw_max = 40;
break;
case 500: if (pw_max > 16) pw_max = 16;
hlfmt_hash (hashlist_format, input_buf, input_len, &hash_buf, &hash_len);
- if (hash_len)
+ bool hash_fmt_error = 0;
+
+ if (hash_len < 1) hash_fmt_error = 1;
+ if (hash_buf == NULL) hash_fmt_error = 1;
+
+ if (hash_fmt_error)
+ {
+ log_info ("WARNING: failed to parse hashes using the '%s' format", strhlfmt (hashlist_format));
+ }
+ else
{
if (opts_type & OPTS_TYPE_HASH_COPY)
{
wpa_t *wpa = (wpa_t *) hashes_buf[hashes_cnt].esalt;
- u8 *pke_ptr = (u8 *) wpa->pke;
-
// do the appending task
snprintf (salt_ptr + cur_pos,
rem_len,
":%02x%02x%02x%02x%02x%02x:%02x%02x%02x%02x%02x%02x",
- pke_ptr[20], pke_ptr[27], pke_ptr[26], pke_ptr[25], pke_ptr[24], pke_ptr[31], // MAC1
- pke_ptr[30], pke_ptr[29], pke_ptr[28], pke_ptr[35], pke_ptr[34], pke_ptr[33]); // MAC2
-
+ wpa->orig_mac1[0],
+ wpa->orig_mac1[1],
+ wpa->orig_mac1[2],
+ wpa->orig_mac1[3],
+ wpa->orig_mac1[4],
+ wpa->orig_mac1[5],
+ wpa->orig_mac2[0],
+ wpa->orig_mac2[1],
+ wpa->orig_mac2[2],
+ wpa->orig_mac2[3],
+ wpa->orig_mac2[4],
+ wpa->orig_mac2[5]);
// memset () the remaining part of the salt
uint line_num = 0;
+ char *line_buf = (char *) mymalloc (HCBUFSIZ);
+
while (!feof (fp))
{
line_num++;
- char line_buf[BUFSIZ] = { 0 };
-
int line_len = fgetl (fp, line_buf);
if (line_len == 0) continue;
hlfmt_hash (hashlist_format, line_buf, line_len, &hash_buf, &hash_len);
+ bool hash_fmt_error = 0;
+
+ if (hash_len < 1) hash_fmt_error = 1;
+ if (hash_buf == NULL) hash_fmt_error = 1;
+
+ if (hash_fmt_error)
+ {
+ log_info ("WARNING: failed to parse hashes using the '%s' format", strhlfmt (hashlist_format));
+
+ continue;
+ }
+
if (username)
{
char *user_buf = NULL;
}
}
+ myfree (line_buf);
+
fclose (fp);
if (data.quiet == 0) log_info_nn ("Parsed Hashes: %u/%u (%0.2f%%)", hashes_avail, hashes_avail, 100.00);
switch (hash_mode)
{
- case 1500: hashes_buf[0].salt->salt_len = 2;
+ case 1500: hashes_buf[0].salt->salt_len = 2;
+ hashes_buf[0].salt->salt_buf[0] = 388; // pure magic
break;
case 1731: hashes_buf[0].salt->salt_len = 4;
break;
((seven_zip_t *) hashes_buf[0].esalt)->data_len = 112;
((seven_zip_t *) hashes_buf[0].esalt)->unpack_size = 112;
break;
+ case 13400: ((keepass_t *) hashes_buf[0].esalt)->version = 2;
+ break;
}
}
break;
case 13000: hashes_buf[0].salt->salt_iter = ROUNDS_RAR5 - 1;
break;
+ case 13200: hashes_buf[0].salt->salt_iter = ROUNDS_AXCRYPT;
+ break;
+ case 13400: hashes_buf[0].salt->salt_iter = ROUNDS_KEEPASS;
+ break;
}
hashes_cnt = 1;
if (fp != NULL)
{
+ char *line_buf = (char *) mymalloc (HCBUFSIZ);
+
+ // to be safe work with a copy (because of line_len loop, i etc)
+ // moved up here because it's easier to handle continue case
+ // it's just 64kb
+
+ char *line_buf_cpy = (char *) mymalloc (HCBUFSIZ);
+
while (!feof (fp))
{
- char line_buf[BUFSIZ] = { 0 };
-
- char *ptr = fgets (line_buf, BUFSIZ - 1, fp);
+ char *ptr = fgets (line_buf, HCBUFSIZ - 1, fp);
if (ptr == NULL) break;
// here we have in line_buf: ESSID:MAC1:MAC2 (without the plain)
// manipulate salt_buf
- // to be safe work with a copy (because of line_len loop, i etc)
-
- char line_buf_cpy[BUFSIZ] = { 0 };
-
memcpy (line_buf_cpy, line_buf, i);
char *mac2_pos = strrchr (line_buf_cpy, ':');
{
wpa_t *wpa = (wpa_t *) found->esalt;
- uint pke[25] = { 0 };
-
- char *pke_ptr = (char *) pke;
-
- for (uint i = 0; i < 25; i++)
- {
- pke[i] = byte_swap_32 (wpa->pke[i]);
- }
-
- u8 mac1[6] = { 0 };
- u8 mac2[6] = { 0 };
-
- memcpy (mac1, pke_ptr + 23, 6);
- memcpy (mac2, pke_ptr + 29, 6);
-
// compare hex string(s) vs binary MAC address(es)
for (uint i = 0, j = 0; i < 6; i++, j += 2)
{
- if (mac1[i] != hex_to_u8 ((const u8 *) &mac1_pos[j]))
+ if (wpa->orig_mac1[i] != hex_to_u8 ((const u8 *) &mac1_pos[j]))
{
found = NULL;
+
break;
}
}
for (uint i = 0, j = 0; i < 6; i++, j += 2)
{
- if (mac2[i] != hex_to_u8 ((const u8 *) &mac2_pos[j]))
+ if (wpa->orig_mac2[i] != hex_to_u8 ((const u8 *) &mac2_pos[j]))
{
found = NULL;
+
break;
}
}
}
}
+ myfree (line_buf_cpy);
+
+ myfree (line_buf);
+
fclose (fp);
}
}
all_kernel_rules_buf = (kernel_rule_t **) mycalloc (rp_files_cnt, sizeof (kernel_rule_t *));
}
- char rule_buf[BUFSIZ] = { 0 };
+ char *rule_buf = (char *) mymalloc (HCBUFSIZ);
int rule_len = 0;
while (!feof (fp))
{
- memset (rule_buf, 0, BUFSIZ);
+ memset (rule_buf, 0, HCBUFSIZ);
rule_len = fgetl (fp, rule_buf);
kernel_rules_avail += INCR_RULES;
}
- memset (rule_buf, 0, BLOCK_SIZE);
+ memset (rule_buf, 0, HCBUFSIZ);
rule_len = (int) generate_random_rule (rule_buf, rp_gen_func_min, rp_gen_func_max);
}
}
+ myfree (rule_buf);
+
/**
* generate NOP rules
*/
return (-1);
}
+
+ if (opencl_platforms_filter != (uint) -1)
+ {
+ uint platform_cnt_mask = ~(((uint) -1 >> platforms_cnt) << platforms_cnt);
+
+ if (opencl_platforms_filter > platform_cnt_mask)
+ {
+ log_error ("ERROR: The platform selected by the --opencl-platforms parameter is larger than the number of available platforms (%d)", platforms_cnt);
+
+ return (-1);
+ }
+ }
}
/**
hc_clGetDeviceIDs (data.ocl, platform, CL_DEVICE_TYPE_ALL, DEVICES_MAX, platform_devices, &platform_devices_cnt);
+ char platform_vendor[INFOSZ] = { 0 };
+
+ hc_clGetPlatformInfo (data.ocl, platform, CL_PLATFORM_VENDOR, sizeof (platform_vendor), platform_vendor, NULL);
+
+ // find our own platform vendor because pocl and mesa are pushing original vendor_id through opencl
+ // this causes trouble with vendor id based macros
+ // we'll assign generic to those without special optimization available
+
+ cl_uint vendor_id = 0;
+
+ if (strcmp (platform_vendor, CL_VENDOR_AMD) == 0)
+ {
+ vendor_id = VENDOR_ID_AMD;
+ }
+ else if (strcmp (platform_vendor, CL_VENDOR_APPLE) == 0)
+ {
+ vendor_id = VENDOR_ID_GENERIC;
+ }
+ else if (strcmp (platform_vendor, CL_VENDOR_INTEL_BEIGNET) == 0)
+ {
+ vendor_id = VENDOR_ID_GENERIC;
+ }
+ else if (strcmp (platform_vendor, CL_VENDOR_INTEL_SDK) == 0)
+ {
+ vendor_id = VENDOR_ID_GENERIC;
+ }
+ else if (strcmp (platform_vendor, CL_VENDOR_MESA) == 0)
+ {
+ vendor_id = VENDOR_ID_GENERIC;
+ }
+ else if (strcmp (platform_vendor, CL_VENDOR_NV) == 0)
+ {
+ vendor_id = VENDOR_ID_NV;
+ }
+ else if (strcmp (platform_vendor, CL_VENDOR_POCL) == 0)
+ {
+ vendor_id = VENDOR_ID_GENERIC;
+ }
+ else
+ {
+ vendor_id = VENDOR_ID_GENERIC;
+ }
+
for (uint platform_devices_id = 0; platform_devices_id < platform_devices_cnt; platform_devices_id++)
{
size_t param_value_size = 0;
hc_device_param_t *device_param = &data.devices_param[device_id];
+ device_param->vendor_id = vendor_id;
+
device_param->device = platform_devices[platform_devices_id];
device_param->device_id = device_id;
device_param->device_type = device_type;
- // vendor_id
-
- cl_uint vendor_id = 0;
-
- hc_clGetDeviceInfo (data.ocl, device_param->device, CL_DEVICE_VENDOR_ID, sizeof (vendor_id), &vendor_id, NULL);
-
- device_param->vendor_id = vendor_id;
-
// device_name
hc_clGetDeviceInfo (data.ocl, device_param->device, CL_DEVICE_NAME, 0, NULL, ¶m_value_size);
// tuning db
- tuning_db_entry_t *tuningdb_entry = tuning_db_search (tuning_db, device_param->device_name, attack_mode, hash_mode);
+ tuning_db_entry_t *tuningdb_entry = tuning_db_search (tuning_db, device_param, attack_mode, hash_mode);
// device_version
myfree (device_opencl_version);
- if (strstr (device_version, "pocl"))
- {
- // pocl returns the real vendor_id in CL_DEVICE_VENDOR_ID which causes many problems because of hms and missing amd_bfe () etc
- // we need to overwrite vendor_id to avoid this. maybe open pocl issue?
-
- cl_uint vendor_id = VENDOR_ID_GENERIC;
-
- device_param->vendor_id = vendor_id;
- }
-
// vector_width
cl_uint vector_width;
if (opencl_vector_width_chgd == 0)
{
- if (tuningdb_entry == NULL)
+ if (tuningdb_entry == NULL || tuningdb_entry->vector_width == -1)
{
if (opti_type & OPTI_TYPE_USES_BITS_64)
{
- hc_clGetDeviceInfo (data.ocl, device_param->device, CL_DEVICE_PREFERRED_VECTOR_WIDTH_LONG, sizeof (vector_width), &vector_width, NULL);
+ hc_clGetDeviceInfo (data.ocl, device_param->device, CL_DEVICE_NATIVE_VECTOR_WIDTH_LONG, sizeof (vector_width), &vector_width, NULL);
}
else
{
- hc_clGetDeviceInfo (data.ocl, device_param->device, CL_DEVICE_PREFERRED_VECTOR_WIDTH_INT, sizeof (vector_width), &vector_width, NULL);
+ hc_clGetDeviceInfo (data.ocl, device_param->device, CL_DEVICE_NATIVE_VECTOR_WIDTH_INT, sizeof (vector_width), &vector_width, NULL);
}
}
else
{
- if (tuningdb_entry->vector_width == -1)
- {
- if (opti_type & OPTI_TYPE_USES_BITS_64)
- {
- hc_clGetDeviceInfo (data.ocl, device_param->device, CL_DEVICE_PREFERRED_VECTOR_WIDTH_LONG, sizeof (vector_width), &vector_width, NULL);
- }
- else
- {
- hc_clGetDeviceInfo (data.ocl, device_param->device, CL_DEVICE_PREFERRED_VECTOR_WIDTH_INT, sizeof (vector_width), &vector_width, NULL);
- }
- }
- else
- {
- vector_width = (cl_uint) tuningdb_entry->vector_width;
- }
+ vector_width = (cl_uint) tuningdb_entry->vector_width;
}
}
else
vector_width = opencl_vector_width;
}
- if (vector_width > 8) vector_width = 8;
+ if (vector_width > 16) vector_width = 16;
device_param->vector_width = vector_width;
if (device_param->skipped == 0)
{
- if (strstr (device_version, "pocl"))
- {
- if (force == 0)
- {
- log_info ("");
- log_info ("ATTENTION! All pocl drivers are known to be broken due to broken LLVM <= 3.7");
- log_info ("You are STRONGLY encouraged not to use it");
- log_info ("You can use --force to override this but do not post error reports if you do so");
- log_info ("");
-
- return (-1);
- }
- }
-
if (device_type & CL_DEVICE_TYPE_GPU)
{
- if (vendor_id == VENDOR_ID_NV)
- {
- if (device_param->kernel_exec_timeout != 0)
- {
- if (data.quiet == 0) log_info ("Device #%u: WARNING! Kernel exec timeout is not disabled, it might cause you errors of code 702", device_id + 1);
- if (data.quiet == 0) log_info (" See the wiki on how to disable it: https://hashcat.net/wiki/doku.php?id=timeout_patch");
- }
- }
- else if (vendor_id == VENDOR_ID_AMD)
+ if (vendor_id == VENDOR_ID_AMD)
{
int catalyst_check = (force == 1) ? 0 : 1;
log_info ("You can use --force to override this but do not post error reports if you do so");
log_info ("");
+ return (-1);
+ }
+ }
+ else if (vendor_id == VENDOR_ID_NV)
+ {
+ if (device_param->kernel_exec_timeout != 0)
+ {
+ if (data.quiet == 0) log_info ("Device #%u: WARNING! Kernel exec timeout is not disabled, it might cause you errors of code 702", device_id + 1);
+ if (data.quiet == 0) log_info (" See the wiki on how to disable it: https://hashcat.net/wiki/doku.php?id=timeout_patch");
+ }
+ }
+ else if (vendor_id == VENDOR_ID_POCL)
+ {
+ if (force == 0)
+ {
+ log_info ("");
+ log_info ("ATTENTION! All pocl drivers are known to be broken due to broken LLVM <= 3.7");
+ log_info ("You are STRONGLY encouraged not to use it");
+ log_info ("You can use --force to override this but do not post error reports if you do so");
+ log_info ("");
+
return (-1);
}
}
* kernel accel and loops tuning db adjustment
*/
- uint _kernel_accel = kernel_accel;
- uint _kernel_loops = kernel_loops;
+ device_param->kernel_accel_min = 1;
+ device_param->kernel_accel_max = 1024;
- tuning_db_entry_t *tuningdb_entry = tuning_db_search (tuning_db, device_param->device_name, attack_mode, hash_mode);
+ device_param->kernel_loops_min = 1;
+ device_param->kernel_loops_max = 1024;
- if (kernel_accel_chgd == 0)
+ tuning_db_entry_t *tuningdb_entry = tuning_db_search (tuning_db, device_param, attack_mode, hash_mode);
+
+ if (tuningdb_entry)
{
- if (tuningdb_entry)
+ u32 _kernel_accel = tuningdb_entry->kernel_accel;
+ u32 _kernel_loops = tuningdb_entry->kernel_loops;
+
+ if (_kernel_accel)
{
- _kernel_accel = tuningdb_entry->kernel_accel;
+ device_param->kernel_accel_min = _kernel_accel;
+ device_param->kernel_accel_max = _kernel_accel;
}
- }
- if (kernel_loops_chgd == 0)
- {
- if (tuningdb_entry)
+ if (_kernel_loops)
{
- _kernel_loops = tuningdb_entry->kernel_loops;
-
if (workload_profile == 1)
{
_kernel_loops = (_kernel_loops > 8) ? _kernel_loops / 8 : 1;
{
_kernel_loops = (_kernel_loops > 4) ? _kernel_loops / 4 : 1;
}
+
+ device_param->kernel_loops_min = _kernel_loops;
+ device_param->kernel_loops_max = _kernel_loops;
}
}
- device_param->kernel_accel = _kernel_accel;
- device_param->kernel_loops = _kernel_loops;
+ // commandline parameters overwrite tuningdb entries
+
+ if (kernel_accel)
+ {
+ device_param->kernel_accel_min = kernel_accel;
+ device_param->kernel_accel_max = kernel_accel;
+ }
+
+ if (kernel_loops)
+ {
+ device_param->kernel_loops_min = kernel_loops;
+ device_param->kernel_loops_max = kernel_loops;
+ }
+
+ /**
+ * activate device
+ */
devices_active++;
}
return (-1);
}
+ // additional check to see if the user has chosen a device that is not within the range of available devices (i.e. larger than devices_cnt)
+
+ if (devices_filter != (uint) -1)
+ {
+ uint devices_cnt_mask = ~(((uint) -1 >> devices_cnt) << devices_cnt);
+
+ if (devices_filter > devices_cnt_mask)
+ {
+ log_error ("ERROR: The device specified by the --opencl-devices parameter is larger than the number of available devices (%d)", devices_cnt);
+
+ return (-1);
+ }
+ }
+
data.devices_cnt = devices_cnt;
data.devices_active = devices_active;
#endif // HAVE_ADK
#endif // HAVE_HWMON
- #ifdef OSX
- if (hash_mode == 3000 || hash_mode == 1500 || hash_mode == 10700)
- {
- if (force == 0)
- {
- log_info ("");
- log_info ("Warning: Hash mode %d is not stable with OSX.", hash_mode);
- log_info ("You can use --force to override this but do not post error reports if you do so");
- log_info ("");
-
- continue;
- }
- }
- #endif
-
#ifdef DEBUG
if (benchmark == 1) log_info ("Hashmode: %d", data.hash_mode);
#endif
* some algorithms need a fixed kernel-loops count
*/
- u32 kernel_loops_min = 1;
- u32 kernel_loops_max = 1024;
-
if (hash_mode == 1500)
{
const u32 kernel_loops_fixed = 1024;
- kernel_loops_min = kernel_loops_fixed;
- kernel_loops_max = kernel_loops_fixed;
+ device_param->kernel_loops_min = kernel_loops_fixed;
+ device_param->kernel_loops_max = kernel_loops_fixed;
}
if (hash_mode == 3000)
{
const u32 kernel_loops_fixed = 1024;
- kernel_loops_min = kernel_loops_fixed;
- kernel_loops_max = kernel_loops_fixed;
+ device_param->kernel_loops_min = kernel_loops_fixed;
+ device_param->kernel_loops_max = kernel_loops_fixed;
}
if (hash_mode == 8900)
{
const u32 kernel_loops_fixed = 1;
- kernel_loops_min = kernel_loops_fixed;
- kernel_loops_max = kernel_loops_fixed;
+ device_param->kernel_loops_min = kernel_loops_fixed;
+ device_param->kernel_loops_max = kernel_loops_fixed;
}
if (hash_mode == 9300)
{
const u32 kernel_loops_fixed = 1;
- kernel_loops_min = kernel_loops_fixed;
- kernel_loops_max = kernel_loops_fixed;
+ device_param->kernel_loops_min = kernel_loops_fixed;
+ device_param->kernel_loops_max = kernel_loops_fixed;
}
if (hash_mode == 12500)
{
const u32 kernel_loops_fixed = ROUNDS_RAR3 / 16;
- kernel_loops_min = kernel_loops_fixed;
- kernel_loops_max = kernel_loops_fixed;
+ device_param->kernel_loops_min = kernel_loops_fixed;
+ device_param->kernel_loops_max = kernel_loops_fixed;
}
+ /**
+ * some algorithms have a maximum kernel-loops count
+ */
+
if (attack_exec == ATTACK_EXEC_OUTSIDE_KERNEL)
{
- if (data.salts_buf[0].salt_iter < kernel_loops_max)
+ if (data.salts_buf[0].salt_iter < device_param->kernel_loops_max)
{
- kernel_loops_max = data.salts_buf[0].salt_iter;
+ device_param->kernel_loops_max = data.salts_buf[0].salt_iter;
}
}
- device_param->kernel_loops_min = kernel_loops_min;
- device_param->kernel_loops_max = kernel_loops_max;
-
- // find out if we would request too much memory on memory blocks which are based on kernel_accel
-
- uint size_pws = 4;
- uint size_tmps = 4;
- uint size_hooks = 4;
-
- uint kernel_accel_min = 1;
- uint kernel_accel_max = 1024;
-
/**
* some algorithms need a special kernel-accel
*/
if (hash_mode == 8900)
{
- kernel_accel_max = 64;
+ device_param->kernel_accel_min = 1;
+ device_param->kernel_accel_max = 64;
}
if (hash_mode == 9300)
{
- kernel_accel_max = 64;
+ device_param->kernel_accel_min = 1;
+ device_param->kernel_accel_max = 64;
}
- while (kernel_accel_max)
+ u32 kernel_accel_min = device_param->kernel_accel_min;
+ u32 kernel_accel_max = device_param->kernel_accel_max;
+
+ // find out if we would request too much memory on memory blocks which are based on kernel_accel
+
+ uint size_pws = 4;
+ uint size_tmps = 4;
+ uint size_hooks = 4;
+
+ while (kernel_accel_max >= kernel_accel_min)
{
uint kernel_power_max = device_processors * kernel_threads * kernel_accel_max;
case 12800: size_tmps = kernel_power_max * sizeof (pbkdf2_sha256_tmp_t); break;
case 12900: size_tmps = kernel_power_max * sizeof (pbkdf2_sha256_tmp_t); break;
case 13000: size_tmps = kernel_power_max * sizeof (pbkdf2_sha256_tmp_t); break;
+ case 13200: size_tmps = kernel_power_max * sizeof (axcrypt_tmp_t); break;
+ case 13400: size_tmps = kernel_power_max * sizeof (keepass_tmp_t); break;
};
// size_hooks
break;
}
+ /*
if (kernel_accel_max == 0)
{
log_error ("Device #%u: Device does not provide enough allocatable device-memory to handle hash-type %u", device_id + 1, data.hash_mode);
return -1;
}
+ */
device_param->kernel_accel_min = kernel_accel_min;
device_param->kernel_accel_max = kernel_accel_max;
+ /*
if (kernel_accel_max < kernel_accel)
{
if (quiet == 0) log_info ("Device #%u: Reduced maximum kernel-accel to %u", device_id + 1, kernel_accel_max);
device_param->kernel_accel = kernel_accel_max;
}
+ */
- const u32 kernel_accel = device_param->kernel_accel;
-
- device_param->size_pws = size_pws;
- device_param->size_tmps = size_tmps;
- device_param->size_hooks = size_hooks;
+ device_param->size_bfs = size_bfs;
+ device_param->size_combs = size_combs;
+ device_param->size_rules = size_rules;
+ device_param->size_rules_c = size_rules_c;
+ device_param->size_pws = size_pws;
+ device_param->size_tmps = size_tmps;
+ device_param->size_hooks = size_hooks;
// do not confuse kernel_accel_max with kernel_accel here
- const u32 kernel_power = device_processors * kernel_threads * kernel_accel;
+ const u32 kernel_power = device_processors * kernel_threads * kernel_accel_max;
device_param->kernel_threads = kernel_threads;
device_param->kernel_power_user = kernel_power;
// we don't have sm_* on vendors not NV but it doesn't matter
- snprintf (build_opts, sizeof (build_opts) - 1, "-I%s/ -DVENDOR_ID=%u -DCUDA_ARCH=%d -DVECT_SIZE=%u -DDEVICE_TYPE=%u", shared_dir, device_param->vendor_id, (device_param->sm_major * 100) + device_param->sm_minor, device_param->vector_width, (u32) device_param->device_type);
+ snprintf (build_opts, sizeof (build_opts) - 1, "-I\"%s/\" -DVENDOR_ID=%u -DCUDA_ARCH=%d -DVECT_SIZE=%u -DDEVICE_TYPE=%u", shared_dir, device_param->vendor_id, (device_param->sm_major * 100) + device_param->sm_minor, device_param->vector_width, (u32) device_param->device_type);
/**
* main kernel
int rc = hc_clBuildProgram (data.ocl, device_param->program, 1, &device_param->device, build_opts, NULL, NULL, false);
+ #ifdef DEBUG
+ size_t build_log_size = 0;
+
+ hc_clGetProgramBuildInfo (data.ocl, device_param->program, device_param->device, CL_PROGRAM_BUILD_LOG, 0, NULL, &build_log_size);
+
+ if (build_log_size > 1)
+ {
+ char *build_log = (char *) malloc (build_log_size + 1);
+
+ memset (build_log, 0, build_log_size + 1);
+
+ hc_clGetProgramBuildInfo (data.ocl, device_param->program, device_param->device, CL_PROGRAM_BUILD_LOG, build_log_size, build_log, NULL);
+
+ puts (build_log);
+
+ free (build_log);
+ }
+ #endif
+
if (rc != 0)
{
device_param->skipped = true;
{
snprintf (build_opts_update, sizeof (build_opts_update) - 1, "%s -DSCRYPT_N=%d -DSCRYPT_R=%d -DSCRYPT_P=%d -DSCRYPT_TMTO=%d", build_opts, data.salts_buf[0].scrypt_N, data.salts_buf[0].scrypt_r, data.salts_buf[0].scrypt_p, 1 << data.salts_buf[0].scrypt_tmto);
}
+ else
+ {
+ snprintf (build_opts_update, sizeof (build_opts_update) - 1, "%s", build_opts);
+ }
int rc = hc_clBuildProgram (data.ocl, device_param->program, 1, &device_param->device, build_opts_update, NULL, NULL, false);
+ #ifdef DEBUG
+ size_t build_log_size = 0;
+
+ hc_clGetProgramBuildInfo (data.ocl, device_param->program, device_param->device, CL_PROGRAM_BUILD_LOG, 0, NULL, &build_log_size);
+
+ if (build_log_size > 1)
+ {
+ char *build_log = (char *) malloc (build_log_size + 1);
+
+ memset (build_log, 0, build_log_size + 1);
+
+ hc_clGetProgramBuildInfo (data.ocl, device_param->program, device_param->device, CL_PROGRAM_BUILD_LOG, build_log_size, build_log, NULL);
+
+ puts (build_log);
+
+ free (build_log);
+ }
+ #endif
+
if (rc != 0)
{
device_param->skipped = true;
+
log_info ("Device #%u: Kernel %s build failure. Proceed without this device.", device_id + 1, source_file);
}
}
device_param->pws_buf = pws_buf;
- pw_cache_t *pw_caches = (pw_cache_t *) mycalloc (64, sizeof (pw_cache_t));
-
- for (int i = 0; i < 64; i++)
- {
- pw_caches[i].pw_buf.pw_len = i;
- pw_caches[i].cnt = 0;
- }
-
- device_param->pw_caches = pw_caches;
-
comb_t *combs_buf = (comb_t *) mycalloc (KERNEL_COMBS, sizeof (comb_t));
device_param->combs_buf = combs_buf;
device_param->hooks_buf = hooks_buf;
- device_param->pw_transpose = pw_transpose_to_hi1;
- device_param->pw_add = pw_add_to_hc1;
-
/**
* kernel args
*/
device_param->kernel_params_amp[5] = &device_param->kernel_params_amp_buf32[5];
device_param->kernel_params_amp[6] = &device_param->kernel_params_amp_buf32[6];
- device_param->kernel_params_tb[0] = &device_param->d_pws_buf;
-
device_param->kernel_params_tm[0] = &device_param->d_bfs_c;
device_param->kernel_params_tm[1] = &device_param->d_tm_c;
{
if (opts_type & OPTS_TYPE_PT_BITSLICE)
{
- snprintf (kernel_name, sizeof (kernel_name) - 1, "m%05d_tb", kern_type);
-
- device_param->kernel_tb = hc_clCreateKernel (data.ocl, device_param->program, kernel_name);
-
snprintf (kernel_name, sizeof (kernel_name) - 1, "m%05d_tm", kern_type);
device_param->kernel_tm = hc_clCreateKernel (data.ocl, device_param->program, kernel_name);
if (opts_type & OPTS_TYPE_PT_BITSLICE)
{
- hc_clSetKernelArg (data.ocl, device_param->kernel_tb, 0, sizeof (cl_mem), device_param->kernel_params_tb[0]);
-
hc_clSetKernelArg (data.ocl, device_param->kernel_tm, 0, sizeof (cl_mem), device_param->kernel_params_tm[0]);
hc_clSetKernelArg (data.ocl, device_param->kernel_tm, 1, sizeof (cl_mem), device_param->kernel_params_tm[1]);
}
if (data.quiet == 0) log_info ("");
/**
- * Inform user which algorithm is checked and at which workload setting
+ * In benchmark-mode, inform user which algorithm is checked
*/
if (benchmark == 1)
return (-1);
}
- char line_buf[BUFSIZ] = { 0 };
+ char *line_buf = (char *) mymalloc (HCBUFSIZ);
while (!feof (mask_fp))
{
- memset (line_buf, 0, BUFSIZ);
+ memset (line_buf, 0, HCBUFSIZ);
int line_len = fgetl (mask_fp, line_buf);
maskcnt++;
}
+ myfree (line_buf);
+
fclose (mask_fp);
}
else
return (-1);
}
- char line_buf[BUFSIZ] = { 0 };
+ char *line_buf = (char *) mymalloc (HCBUFSIZ);
uint masks_avail = 1;
while (!feof (mask_fp))
{
- memset (line_buf, 0, BUFSIZ);
+ memset (line_buf, 0, HCBUFSIZ);
int line_len = fgetl (mask_fp, line_buf);
maskcnt++;
}
+ myfree (line_buf);
+
fclose (mask_fp);
mask_from_file = 1;
return (-1);
}
- char line_buf[BUFSIZ] = { 0 };
+ char *line_buf = (char *) mymalloc (HCBUFSIZ);
uint masks_avail = 1;
while (!feof (mask_fp))
{
- memset (line_buf, 0, BUFSIZ);
+ memset (line_buf, 0, HCBUFSIZ);
int line_len = fgetl (mask_fp, line_buf);
maskcnt++;
}
+ myfree (line_buf);
+
fclose (mask_fp);
mask_from_file = 1;
device_param->speed_pos = 0;
memset (device_param->speed_cnt, 0, SPEED_CACHE * sizeof (u64));
- memset (device_param->speed_ms, 0, SPEED_CACHE * sizeof (float));
- memset (device_param->speed_rec, 0, SPEED_CACHE * sizeof (hc_timer_t));
+ memset (device_param->speed_ms, 0, SPEED_CACHE * sizeof (double));
device_param->exec_pos = 0;
// some more resets:
- if (device_param->pw_caches) memset (device_param->pw_caches, 0, 64 * sizeof (pw_cache_t));
-
if (device_param->pws_buf) memset (device_param->pws_buf, 0, device_param->size_pws);
- device_param->pw_cnt = 0;
device_param->pws_cnt = 0;
device_param->words_off = 0;
local_free (device_param->result);
- local_free (device_param->pw_caches);
-
local_free (device_param->combs_buf);
local_free (device_param->hooks_buf);
if (device_param->kernel_mp) hc_clReleaseKernel (data.ocl, device_param->kernel_mp);
if (device_param->kernel_mp_l) hc_clReleaseKernel (data.ocl, device_param->kernel_mp_l);
if (device_param->kernel_mp_r) hc_clReleaseKernel (data.ocl, device_param->kernel_mp_r);
- if (device_param->kernel_tb) hc_clReleaseKernel (data.ocl, device_param->kernel_tb);
if (device_param->kernel_tm) hc_clReleaseKernel (data.ocl, device_param->kernel_tm);
if (device_param->kernel_amp) hc_clReleaseKernel (data.ocl, device_param->kernel_amp);