/** * Author......: See docs/credits.txt * License.....: MIT */ #include "common.h" #include "types.h" #include "event.h" #include "opencl.h" #include "status.h" #include "terminal.h" #include "autotune.h" static double try_run (hashcat_ctx_t *hashcat_ctx, hc_device_param_t *device_param, const u32 kernel_accel, const u32 kernel_loops) { hashconfig_t *hashconfig = hashcat_ctx->hashconfig; device_param->kernel_params_buf32[28] = 0; device_param->kernel_params_buf32[29] = kernel_loops; // not a bug, both need to be set device_param->kernel_params_buf32[30] = kernel_loops; // because there's two variables for inner iters for slow and fast hashes if (hashconfig->attack_exec == ATTACK_EXEC_INSIDE_KERNEL) { const u32 kernel_power_try = device_param->device_processors * device_param->kernel_threads_by_wgs_kernel1 * kernel_accel; run_kernel (hashcat_ctx, device_param, KERN_RUN_1, kernel_power_try, true, 0); } else { const u32 kernel_power_try = device_param->device_processors * device_param->kernel_threads_by_wgs_kernel2 * kernel_accel; run_kernel (hashcat_ctx, device_param, KERN_RUN_2, kernel_power_try, true, 0); } const double exec_msec_prev = get_avg_exec_time (device_param, 1); return exec_msec_prev; } static int autotune (hashcat_ctx_t *hashcat_ctx, hc_device_param_t *device_param) { hashconfig_t *hashconfig = hashcat_ctx->hashconfig; opencl_ctx_t *opencl_ctx = hashcat_ctx->opencl_ctx; straight_ctx_t *straight_ctx = hashcat_ctx->straight_ctx; user_options_extra_t *user_options_extra = hashcat_ctx->user_options_extra; const double target_msec = opencl_ctx->target_msec; const u32 kernel_accel_min = device_param->kernel_accel_min; const u32 kernel_accel_max = device_param->kernel_accel_max; const u32 kernel_loops_min = device_param->kernel_loops_min; const u32 kernel_loops_max = device_param->kernel_loops_max; u32 kernel_accel = kernel_accel_min; u32 kernel_loops = kernel_loops_min; // in this case the user specified a fixed -n and -u on the commandline // no way to tune anything // but we need to run a few caching rounds if ((kernel_accel_min == kernel_accel_max) && (kernel_loops_min == kernel_loops_max)) { #if defined (DEBUG) // don't do any autotune in debug mode in this case // we're propably during kernel development #else if (hashconfig->hash_mode != 2000) { try_run (hashcat_ctx, device_param, kernel_accel, kernel_loops); try_run (hashcat_ctx, device_param, kernel_accel, kernel_loops); try_run (hashcat_ctx, device_param, kernel_accel, kernel_loops); try_run (hashcat_ctx, device_param, kernel_accel, kernel_loops); } #endif device_param->kernel_accel = kernel_accel; device_param->kernel_loops = kernel_loops; const u32 kernel_power = device_param->device_processors * device_param->kernel_threads_by_user * device_param->kernel_accel; device_param->kernel_power = kernel_power; return 0; } // from here it's clear we are allowed to autotune // so let's init some fake words const u32 kernel_power_max = device_param->device_processors * device_param->kernel_threads_by_user * kernel_accel_max; int CL_rc; if (user_options_extra->attack_kern == ATTACK_KERN_BF) { CL_rc = run_kernel_memset (hashcat_ctx, device_param, device_param->d_pws_buf, 7, kernel_power_max * sizeof (pw_t)); if (CL_rc == -1) return -1; } else { for (u32 i = 0; i < kernel_power_max; i++) { device_param->pws_buf[i].i[0] = i; device_param->pws_buf[i].i[1] = 0x01234567; device_param->pws_buf[i].pw_len = 7 + (i & 7); } CL_rc = hc_clEnqueueWriteBuffer (hashcat_ctx, device_param->command_queue, device_param->d_pws_buf, CL_TRUE, 0, kernel_power_max * sizeof (pw_t), device_param->pws_buf, 0, NULL, NULL); if (CL_rc == -1) return -1; } if (hashconfig->attack_exec == ATTACK_EXEC_INSIDE_KERNEL) { if (straight_ctx->kernel_rules_cnt > 1) { CL_rc = hc_clEnqueueCopyBuffer (hashcat_ctx, device_param->command_queue, device_param->d_rules, device_param->d_rules_c, 0, 0, MIN (kernel_loops_max, KERNEL_RULES) * sizeof (kernel_rule_t), 0, NULL, NULL); if (CL_rc == -1) return -1; } } else { CL_rc = run_kernel_amp (hashcat_ctx, device_param, kernel_power_max); if (CL_rc == -1) return -1; } #define VERIFIER_CNT 1 // first find out highest kernel-loops that stays below target_msec if (kernel_loops_min < kernel_loops_max) { for (kernel_loops = kernel_loops_max; kernel_loops > kernel_loops_min; kernel_loops >>= 1) { double exec_msec = try_run (hashcat_ctx, device_param, kernel_accel_min, kernel_loops); for (int i = 0; i < VERIFIER_CNT; i++) { double exec_msec_v = try_run (hashcat_ctx, device_param, kernel_accel_min, kernel_loops); exec_msec = MIN (exec_msec, exec_msec_v); } if (exec_msec < target_msec) break; } } // now the same for kernel-accel but with the new kernel-loops from previous loop set #define STEPS_CNT 10 if (kernel_accel_min < kernel_accel_max) { for (int i = 0; i < STEPS_CNT; i++) { const u32 kernel_accel_try = 1u << i; if (kernel_accel_try < kernel_accel_min) continue; if (kernel_accel_try > kernel_accel_max) break; double exec_msec = try_run (hashcat_ctx, device_param, kernel_accel_try, kernel_loops); for (int verifier_idx = 0; verifier_idx < VERIFIER_CNT; verifier_idx++) { double exec_msec_v = try_run (hashcat_ctx, device_param, kernel_accel_try, kernel_loops); exec_msec = MIN (exec_msec, exec_msec_v); } if (exec_msec > target_msec) break; kernel_accel = kernel_accel_try; } } // at this point we want to know the actual runtime for the following reason: // we need a reference for the balancing loop following up, and this // the balancing loop can have an effect that the creates a new opportunity, for example: // if the target is 95 ms and the current runtime is 48ms the above loop // stopped the execution because the previous exec_msec was > 95ms // due to the rebalance it's possible that the runtime reduces from 48ms to 47ms // and this creates the possibility to double the workload -> 47 * 2 = 95ms, which is < 96ms double exec_msec_pre_final = try_run (hashcat_ctx, device_param, kernel_accel, kernel_loops); for (int verifier_idx = 0; verifier_idx < VERIFIER_CNT; verifier_idx++) { double exec_msec_pre_final_v = try_run (hashcat_ctx, device_param, kernel_accel, kernel_loops); exec_msec_pre_final = MIN (exec_msec_pre_final, exec_msec_pre_final_v); } u32 diff = kernel_loops - kernel_accel; if ((kernel_loops_min < kernel_loops_max) && (kernel_accel_min < kernel_accel_max)) { u32 kernel_accel_orig = kernel_accel; u32 kernel_loops_orig = kernel_loops; for (u32 f = 1; f < 1024; f++) { const u32 kernel_accel_try = kernel_accel_orig * f; const u32 kernel_loops_try = kernel_loops_orig / f; if (kernel_accel_try > kernel_accel_max) break; if (kernel_loops_try < kernel_loops_min) break; u32 diff_new = kernel_loops_try - kernel_accel_try; if (diff_new > diff) break; diff_new = diff; double exec_msec = try_run (hashcat_ctx, device_param, kernel_accel_try, kernel_loops_try); for (int verifier_idx = 0; verifier_idx < VERIFIER_CNT; verifier_idx++) { double exec_msec_v = try_run (hashcat_ctx, device_param, kernel_accel_try, kernel_loops_try); exec_msec = MIN (exec_msec, exec_msec_v); } for (int verifier_idx = 0; verifier_idx < VERIFIER_CNT; verifier_idx++) { exec_msec_pre_final = exec_msec; kernel_accel = kernel_accel_try; kernel_loops = kernel_loops_try; } } } const double exec_left = target_msec / exec_msec_pre_final; const double accel_left = kernel_accel_max / kernel_accel; const double exec_accel_min = MIN (exec_left, accel_left); // we want that to be int if (exec_accel_min >= 1.0) { // this is safe to not overflow kernel_accel_max because of accel_left kernel_accel *= (u32) exec_accel_min; } // reset them fake words /* memset (device_param->pws_buf, 0, kernel_power_max * sizeof (pw_t)); hc_clEnqueueWriteBuffer (hashcat_ctx, device_param->command_queue, device_param->d_pws_buf, CL_TRUE, 0, kernel_power_max * sizeof (pw_t), device_param->pws_buf, 0, NULL, NULL); hc_clEnqueueWriteBuffer (hashcat_ctx, device_param->command_queue, device_param->d_pws_amp_buf, CL_TRUE, 0, kernel_power_max * sizeof (pw_t), device_param->pws_buf, 0, NULL, NULL); */ CL_rc = run_kernel_memset (hashcat_ctx, device_param, device_param->d_pws_buf, 0, kernel_power_max * sizeof (pw_t)); if (CL_rc == -1) return -1; if (hashconfig->attack_exec == ATTACK_EXEC_OUTSIDE_KERNEL) { CL_rc = run_kernel_memset (hashcat_ctx, device_param, device_param->d_pws_amp_buf, 0, kernel_power_max * sizeof (pw_t)); if (CL_rc == -1) return -1; } // reset timer device_param->exec_pos = 0; memset (device_param->exec_msec, 0, EXEC_CACHE * sizeof (double)); memset (device_param->exec_us_prev1, 0, EXPECTED_ITERATIONS * sizeof (double)); memset (device_param->exec_us_prev2, 0, EXPECTED_ITERATIONS * sizeof (double)); memset (device_param->exec_us_prev3, 0, EXPECTED_ITERATIONS * sizeof (double)); // store device_param->kernel_accel = kernel_accel; device_param->kernel_loops = kernel_loops; const u32 kernel_power = device_param->device_processors * device_param->kernel_threads_by_user * device_param->kernel_accel; device_param->kernel_power = kernel_power; #if defined (DEBUG) user_options_t *user_options = hashcat_ctx->user_options; if (user_options->quiet == false) { clear_prompt (); printf ( "- Device #%u: autotuned kernel-accel to %u" EOL "- Device #%u: autotuned kernel-loops to %u" EOL, device_param->device_id + 1, kernel_accel, device_param->device_id + 1, kernel_loops ); send_prompt (); } #endif return 0; } void *thread_autotune (void *p) { thread_param_t *thread_param = (thread_param_t *) p; hashcat_ctx_t *hashcat_ctx = thread_param->hashcat_ctx; opencl_ctx_t *opencl_ctx = hashcat_ctx->opencl_ctx; if (opencl_ctx->enabled == false) return NULL; hc_device_param_t *device_param = opencl_ctx->devices_param + thread_param->tid; if (device_param->skipped == true) return NULL; const int rc_autotune = autotune (hashcat_ctx, device_param); if (rc_autotune == -1) { // we should do something here, tell hashcat main that autotune failed to abort } return NULL; }