1
mirror of https://github.com/hashcat/hashcat synced 2024-07-20 16:24:12 +02:00
hashcat/OpenCL/m06223.cl
jsteube dad03e394d Fixed two major problems
1) SIMD code for all attack-mode

Macro vector_accessible() was not refactored and missing completely.
Had to rename variables rules_cnt, combs_cnt and bfs_cnt into il_cnt which was a good thing anyway as with new SIMD code they all act in the same way.

2) SIMD code for attack-mode 0

With new SIMD code, apply_rules_vect() has to return u32 not u32x.
This has massive impact on all *_a0 kernels.

I've rewritten most of them. Deep testing using test.sh is still required.

Some kernel need more fixes:

- Some are kind of completely incompatible like m10400 but they still use old check_* includes, we should get rid of them as they are no longer neccessary as we have simd.c
- Some have a chance but require additional effort like m11500. We can use commented out "#define NEW_SIMD_CODE" to find them

This change can have negative impact on -a0 performance for device that require vectorization. That is mostly CPU devices. New GPU's are all scalar, so they wont get hurt by this.
This change also proofes that there's no way to efficiently vectorize kernel rules with new SIMD code, but it enables the addition of the rule functions like @ that we were missing for some long time. This is a TODO.
2016-02-27 17:18:54 +01:00

747 lines
24 KiB
Common Lisp

/**
* Author......: Jens Steube <jens.steube@gmail.com>
* License.....: MIT
*/
#define _SHA512_
#include "include/constants.h"
#include "include/kernel_vendor.h"
#define DGST_R0 0
#define DGST_R1 1
#define DGST_R2 2
#define DGST_R3 3
#include "include/kernel_functions.c"
#include "OpenCL/types_ocl.c"
#include "OpenCL/common.c"
#include "OpenCL/kernel_aes256.c"
#include "OpenCL/kernel_twofish256.c"
#include "OpenCL/kernel_serpent256.c"
__constant u64 k_sha512[80] =
{
SHA512C00, SHA512C01, SHA512C02, SHA512C03,
SHA512C04, SHA512C05, SHA512C06, SHA512C07,
SHA512C08, SHA512C09, SHA512C0a, SHA512C0b,
SHA512C0c, SHA512C0d, SHA512C0e, SHA512C0f,
SHA512C10, SHA512C11, SHA512C12, SHA512C13,
SHA512C14, SHA512C15, SHA512C16, SHA512C17,
SHA512C18, SHA512C19, SHA512C1a, SHA512C1b,
SHA512C1c, SHA512C1d, SHA512C1e, SHA512C1f,
SHA512C20, SHA512C21, SHA512C22, SHA512C23,
SHA512C24, SHA512C25, SHA512C26, SHA512C27,
SHA512C28, SHA512C29, SHA512C2a, SHA512C2b,
SHA512C2c, SHA512C2d, SHA512C2e, SHA512C2f,
SHA512C30, SHA512C31, SHA512C32, SHA512C33,
SHA512C34, SHA512C35, SHA512C36, SHA512C37,
SHA512C38, SHA512C39, SHA512C3a, SHA512C3b,
SHA512C3c, SHA512C3d, SHA512C3e, SHA512C3f,
SHA512C40, SHA512C41, SHA512C42, SHA512C43,
SHA512C44, SHA512C45, SHA512C46, SHA512C47,
SHA512C48, SHA512C49, SHA512C4a, SHA512C4b,
SHA512C4c, SHA512C4d, SHA512C4e, SHA512C4f,
};
static void sha512_transform (const u64 w[16], u64 dgst[8])
{
u64 a = dgst[0];
u64 b = dgst[1];
u64 c = dgst[2];
u64 d = dgst[3];
u64 e = dgst[4];
u64 f = dgst[5];
u64 g = dgst[6];
u64 h = dgst[7];
u64 w0_t = w[ 0];
u64 w1_t = w[ 1];
u64 w2_t = w[ 2];
u64 w3_t = w[ 3];
u64 w4_t = w[ 4];
u64 w5_t = w[ 5];
u64 w6_t = w[ 6];
u64 w7_t = w[ 7];
u64 w8_t = w[ 8];
u64 w9_t = w[ 9];
u64 wa_t = w[10];
u64 wb_t = w[11];
u64 wc_t = w[12];
u64 wd_t = w[13];
u64 we_t = w[14];
u64 wf_t = w[15];
#define ROUND_EXPAND() \
{ \
w0_t = SHA512_EXPAND (we_t, w9_t, w1_t, w0_t); \
w1_t = SHA512_EXPAND (wf_t, wa_t, w2_t, w1_t); \
w2_t = SHA512_EXPAND (w0_t, wb_t, w3_t, w2_t); \
w3_t = SHA512_EXPAND (w1_t, wc_t, w4_t, w3_t); \
w4_t = SHA512_EXPAND (w2_t, wd_t, w5_t, w4_t); \
w5_t = SHA512_EXPAND (w3_t, we_t, w6_t, w5_t); \
w6_t = SHA512_EXPAND (w4_t, wf_t, w7_t, w6_t); \
w7_t = SHA512_EXPAND (w5_t, w0_t, w8_t, w7_t); \
w8_t = SHA512_EXPAND (w6_t, w1_t, w9_t, w8_t); \
w9_t = SHA512_EXPAND (w7_t, w2_t, wa_t, w9_t); \
wa_t = SHA512_EXPAND (w8_t, w3_t, wb_t, wa_t); \
wb_t = SHA512_EXPAND (w9_t, w4_t, wc_t, wb_t); \
wc_t = SHA512_EXPAND (wa_t, w5_t, wd_t, wc_t); \
wd_t = SHA512_EXPAND (wb_t, w6_t, we_t, wd_t); \
we_t = SHA512_EXPAND (wc_t, w7_t, wf_t, we_t); \
wf_t = SHA512_EXPAND (wd_t, w8_t, w0_t, wf_t); \
}
#define ROUND_STEP(i) \
{ \
SHA512_STEP (SHA512_F0o, SHA512_F1o, a, b, c, d, e, f, g, h, w0_t, k_sha512[i + 0]); \
SHA512_STEP (SHA512_F0o, SHA512_F1o, h, a, b, c, d, e, f, g, w1_t, k_sha512[i + 1]); \
SHA512_STEP (SHA512_F0o, SHA512_F1o, g, h, a, b, c, d, e, f, w2_t, k_sha512[i + 2]); \
SHA512_STEP (SHA512_F0o, SHA512_F1o, f, g, h, a, b, c, d, e, w3_t, k_sha512[i + 3]); \
SHA512_STEP (SHA512_F0o, SHA512_F1o, e, f, g, h, a, b, c, d, w4_t, k_sha512[i + 4]); \
SHA512_STEP (SHA512_F0o, SHA512_F1o, d, e, f, g, h, a, b, c, w5_t, k_sha512[i + 5]); \
SHA512_STEP (SHA512_F0o, SHA512_F1o, c, d, e, f, g, h, a, b, w6_t, k_sha512[i + 6]); \
SHA512_STEP (SHA512_F0o, SHA512_F1o, b, c, d, e, f, g, h, a, w7_t, k_sha512[i + 7]); \
SHA512_STEP (SHA512_F0o, SHA512_F1o, a, b, c, d, e, f, g, h, w8_t, k_sha512[i + 8]); \
SHA512_STEP (SHA512_F0o, SHA512_F1o, h, a, b, c, d, e, f, g, w9_t, k_sha512[i + 9]); \
SHA512_STEP (SHA512_F0o, SHA512_F1o, g, h, a, b, c, d, e, f, wa_t, k_sha512[i + 10]); \
SHA512_STEP (SHA512_F0o, SHA512_F1o, f, g, h, a, b, c, d, e, wb_t, k_sha512[i + 11]); \
SHA512_STEP (SHA512_F0o, SHA512_F1o, e, f, g, h, a, b, c, d, wc_t, k_sha512[i + 12]); \
SHA512_STEP (SHA512_F0o, SHA512_F1o, d, e, f, g, h, a, b, c, wd_t, k_sha512[i + 13]); \
SHA512_STEP (SHA512_F0o, SHA512_F1o, c, d, e, f, g, h, a, b, we_t, k_sha512[i + 14]); \
SHA512_STEP (SHA512_F0o, SHA512_F1o, b, c, d, e, f, g, h, a, wf_t, k_sha512[i + 15]); \
}
ROUND_STEP (0);
//#pragma unroll
for (int i = 16; i < 80; i += 16)
{
ROUND_EXPAND (); ROUND_STEP (i);
}
dgst[0] += a;
dgst[1] += b;
dgst[2] += c;
dgst[3] += d;
dgst[4] += e;
dgst[5] += f;
dgst[6] += g;
dgst[7] += h;
}
static void hmac_run (const u64 w1[16], const u64 ipad[8], const u64 opad[8], u64 dgst[8])
{
dgst[0] = ipad[0];
dgst[1] = ipad[1];
dgst[2] = ipad[2];
dgst[3] = ipad[3];
dgst[4] = ipad[4];
dgst[5] = ipad[5];
dgst[6] = ipad[6];
dgst[7] = ipad[7];
sha512_transform (w1, dgst);
u64 w[16];
w[ 0] = dgst[0];
w[ 1] = dgst[1];
w[ 2] = dgst[2];
w[ 3] = dgst[3];
w[ 4] = dgst[4];
w[ 5] = dgst[5];
w[ 6] = dgst[6];
w[ 7] = dgst[7];
w[ 8] = 0x8000000000000000;
w[ 9] = 0;
w[10] = 0;
w[11] = 0;
w[12] = 0;
w[13] = 0;
w[14] = 0;
w[15] = (128 + 64) * 8;
dgst[0] = opad[0];
dgst[1] = opad[1];
dgst[2] = opad[2];
dgst[3] = opad[3];
dgst[4] = opad[4];
dgst[5] = opad[5];
dgst[6] = opad[6];
dgst[7] = opad[7];
sha512_transform (w, dgst);
}
static void hmac_init (u64 w[16], u64 ipad[8], u64 opad[8])
{
w[ 0] ^= 0x3636363636363636;
w[ 1] ^= 0x3636363636363636;
w[ 2] ^= 0x3636363636363636;
w[ 3] ^= 0x3636363636363636;
w[ 4] ^= 0x3636363636363636;
w[ 5] ^= 0x3636363636363636;
w[ 6] ^= 0x3636363636363636;
w[ 7] ^= 0x3636363636363636;
w[ 8] ^= 0x3636363636363636;
w[ 9] ^= 0x3636363636363636;
w[10] ^= 0x3636363636363636;
w[11] ^= 0x3636363636363636;
w[12] ^= 0x3636363636363636;
w[13] ^= 0x3636363636363636;
w[14] ^= 0x3636363636363636;
w[15] ^= 0x3636363636363636;
ipad[0] = SHA512M_A;
ipad[1] = SHA512M_B;
ipad[2] = SHA512M_C;
ipad[3] = SHA512M_D;
ipad[4] = SHA512M_E;
ipad[5] = SHA512M_F;
ipad[6] = SHA512M_G;
ipad[7] = SHA512M_H;
sha512_transform (w, ipad);
w[ 0] ^= 0x6a6a6a6a6a6a6a6a;
w[ 1] ^= 0x6a6a6a6a6a6a6a6a;
w[ 2] ^= 0x6a6a6a6a6a6a6a6a;
w[ 3] ^= 0x6a6a6a6a6a6a6a6a;
w[ 4] ^= 0x6a6a6a6a6a6a6a6a;
w[ 5] ^= 0x6a6a6a6a6a6a6a6a;
w[ 6] ^= 0x6a6a6a6a6a6a6a6a;
w[ 7] ^= 0x6a6a6a6a6a6a6a6a;
w[ 8] ^= 0x6a6a6a6a6a6a6a6a;
w[ 9] ^= 0x6a6a6a6a6a6a6a6a;
w[10] ^= 0x6a6a6a6a6a6a6a6a;
w[11] ^= 0x6a6a6a6a6a6a6a6a;
w[12] ^= 0x6a6a6a6a6a6a6a6a;
w[13] ^= 0x6a6a6a6a6a6a6a6a;
w[14] ^= 0x6a6a6a6a6a6a6a6a;
w[15] ^= 0x6a6a6a6a6a6a6a6a;
opad[0] = SHA512M_A;
opad[1] = SHA512M_B;
opad[2] = SHA512M_C;
opad[3] = SHA512M_D;
opad[4] = SHA512M_E;
opad[5] = SHA512M_F;
opad[6] = SHA512M_G;
opad[7] = SHA512M_H;
sha512_transform (w, opad);
}
static u32 u8add (const u32 a, const u32 b)
{
const u32 a1 = (a >> 0) & 0xff;
const u32 a2 = (a >> 8) & 0xff;
const u32 a3 = (a >> 16) & 0xff;
const u32 a4 = (a >> 24) & 0xff;
const u32 b1 = (b >> 0) & 0xff;
const u32 b2 = (b >> 8) & 0xff;
const u32 b3 = (b >> 16) & 0xff;
const u32 b4 = (b >> 24) & 0xff;
const u32 r1 = (a1 + b1) & 0xff;
const u32 r2 = (a2 + b2) & 0xff;
const u32 r3 = (a3 + b3) & 0xff;
const u32 r4 = (a4 + b4) & 0xff;
const u32 r = r1 << 0
| r2 << 8
| r3 << 16
| r4 << 24;
return r;
}
__kernel void m06223_init (__global pw_t *pws, __global kernel_rule_t *rules_buf, __global comb_t *combs_buf, __global bf_t *bfs_buf, __global tc64_tmp_t *tmps, __global void *hooks, __global u32 *bitmaps_buf_s1_a, __global u32 *bitmaps_buf_s1_b, __global u32 *bitmaps_buf_s1_c, __global u32 *bitmaps_buf_s1_d, __global u32 *bitmaps_buf_s2_a, __global u32 *bitmaps_buf_s2_b, __global u32 *bitmaps_buf_s2_c, __global u32 *bitmaps_buf_s2_d, __global plain_t *plains_buf, __global digest_t *digests_buf, __global u32 *hashes_shown, __global salt_t *salt_bufs, __global tc_t *esalt_bufs, __global u32 *d_return_buf, __global u32 *d_scryptV_buf, const u32 bitmap_mask, const u32 bitmap_shift1, const u32 bitmap_shift2, const u32 salt_pos, const u32 loop_pos, const u32 loop_cnt, const u32 il_cnt, const u32 digests_cnt, const u32 digests_offset, const u32 combs_mode, const u32 gid_max)
{
/**
* base
*/
const u32 gid = get_global_id (0);
if (gid >= gid_max) return;
u32 w0[4];
w0[0] = pws[gid].i[ 0];
w0[1] = pws[gid].i[ 1];
w0[2] = pws[gid].i[ 2];
w0[3] = pws[gid].i[ 3];
u32 w1[4];
w1[0] = pws[gid].i[ 4];
w1[1] = pws[gid].i[ 5];
w1[2] = pws[gid].i[ 6];
w1[3] = pws[gid].i[ 7];
u32 w2[4];
w2[0] = pws[gid].i[ 8];
w2[1] = pws[gid].i[ 9];
w2[2] = pws[gid].i[10];
w2[3] = pws[gid].i[11];
u32 w3[4];
w3[0] = pws[gid].i[12];
w3[1] = pws[gid].i[13];
w3[2] = pws[gid].i[14];
w3[3] = pws[gid].i[15];
/**
* keyfile
*/
w0[0] = u8add (w0[0], esalt_bufs[salt_pos].keyfile_buf[ 0]);
w0[1] = u8add (w0[1], esalt_bufs[salt_pos].keyfile_buf[ 1]);
w0[2] = u8add (w0[2], esalt_bufs[salt_pos].keyfile_buf[ 2]);
w0[3] = u8add (w0[3], esalt_bufs[salt_pos].keyfile_buf[ 3]);
w1[0] = u8add (w1[0], esalt_bufs[salt_pos].keyfile_buf[ 4]);
w1[1] = u8add (w1[1], esalt_bufs[salt_pos].keyfile_buf[ 5]);
w1[2] = u8add (w1[2], esalt_bufs[salt_pos].keyfile_buf[ 6]);
w1[3] = u8add (w1[3], esalt_bufs[salt_pos].keyfile_buf[ 7]);
w2[0] = u8add (w2[0], esalt_bufs[salt_pos].keyfile_buf[ 8]);
w2[1] = u8add (w2[1], esalt_bufs[salt_pos].keyfile_buf[ 9]);
w2[2] = u8add (w2[2], esalt_bufs[salt_pos].keyfile_buf[10]);
w2[3] = u8add (w2[3], esalt_bufs[salt_pos].keyfile_buf[11]);
w3[0] = u8add (w3[0], esalt_bufs[salt_pos].keyfile_buf[12]);
w3[1] = u8add (w3[1], esalt_bufs[salt_pos].keyfile_buf[13]);
w3[2] = u8add (w3[2], esalt_bufs[salt_pos].keyfile_buf[14]);
w3[3] = u8add (w3[3], esalt_bufs[salt_pos].keyfile_buf[15]);
/**
* salt
*/
u64 salt_buf[16];
// swap fehlt
salt_buf[ 0] = ((u64) swap32 (esalt_bufs[salt_pos].salt_buf[ 0])) << 32 | (u64) swap32 (esalt_bufs[salt_pos].salt_buf[ 1]);
salt_buf[ 1] = ((u64) swap32 (esalt_bufs[salt_pos].salt_buf[ 2])) << 32 | (u64) swap32 (esalt_bufs[salt_pos].salt_buf[ 3]);
salt_buf[ 2] = ((u64) swap32 (esalt_bufs[salt_pos].salt_buf[ 4])) << 32 | (u64) swap32 (esalt_bufs[salt_pos].salt_buf[ 5]);
salt_buf[ 3] = ((u64) swap32 (esalt_bufs[salt_pos].salt_buf[ 6])) << 32 | (u64) swap32 (esalt_bufs[salt_pos].salt_buf[ 7]);
salt_buf[ 4] = ((u64) swap32 (esalt_bufs[salt_pos].salt_buf[ 8])) << 32 | (u64) swap32 (esalt_bufs[salt_pos].salt_buf[ 9]);
salt_buf[ 5] = ((u64) swap32 (esalt_bufs[salt_pos].salt_buf[10])) << 32 | (u64) swap32 (esalt_bufs[salt_pos].salt_buf[11]);
salt_buf[ 6] = ((u64) swap32 (esalt_bufs[salt_pos].salt_buf[12])) << 32 | (u64) swap32 (esalt_bufs[salt_pos].salt_buf[13]);
salt_buf[ 7] = ((u64) swap32 (esalt_bufs[salt_pos].salt_buf[14])) << 32 | (u64) swap32 (esalt_bufs[salt_pos].salt_buf[15]);
salt_buf[ 8] = 0;
salt_buf[ 9] = 0;
salt_buf[10] = 0;
salt_buf[11] = 0;
salt_buf[12] = 0;
salt_buf[13] = 0;
salt_buf[14] = 0;
salt_buf[15] = (128 + 64 + 4) * 8;
const u32 truecrypt_mdlen = salt_bufs[0].truecrypt_mdlen;
u64 w[16];
w[ 0] = ((u64) swap32 (w0[0])) << 32 | (u64) swap32 (w0[1]);
w[ 1] = ((u64) swap32 (w0[2])) << 32 | (u64) swap32 (w0[3]);
w[ 2] = ((u64) swap32 (w1[0])) << 32 | (u64) swap32 (w1[1]);
w[ 3] = ((u64) swap32 (w1[2])) << 32 | (u64) swap32 (w1[3]);
w[ 4] = ((u64) swap32 (w2[0])) << 32 | (u64) swap32 (w2[1]);
w[ 5] = ((u64) swap32 (w2[2])) << 32 | (u64) swap32 (w2[3]);
w[ 6] = ((u64) swap32 (w3[0])) << 32 | (u64) swap32 (w3[1]);
w[ 7] = ((u64) swap32 (w3[2])) << 32 | (u64) swap32 (w3[3]);
w[ 8] = 0;
w[ 9] = 0;
w[10] = 0;
w[11] = 0;
w[12] = 0;
w[13] = 0;
w[14] = 0;
w[15] = 0;
u64 ipad[8];
u64 opad[8];
hmac_init (w, ipad, opad);
tmps[gid].ipad[0] = ipad[0];
tmps[gid].ipad[1] = ipad[1];
tmps[gid].ipad[2] = ipad[2];
tmps[gid].ipad[3] = ipad[3];
tmps[gid].ipad[4] = ipad[4];
tmps[gid].ipad[5] = ipad[5];
tmps[gid].ipad[6] = ipad[6];
tmps[gid].ipad[7] = ipad[7];
tmps[gid].opad[0] = opad[0];
tmps[gid].opad[1] = opad[1];
tmps[gid].opad[2] = opad[2];
tmps[gid].opad[3] = opad[3];
tmps[gid].opad[4] = opad[4];
tmps[gid].opad[5] = opad[5];
tmps[gid].opad[6] = opad[6];
tmps[gid].opad[7] = opad[7];
for (u32 i = 0, j = 1; i < (truecrypt_mdlen / 8 / 8); i += 8, j += 1)
{
salt_buf[8] = (u64) j << 32 | (u64) 0x80000000;
u64 dgst[8];
hmac_run (salt_buf, ipad, opad, dgst);
tmps[gid].dgst[i + 0] = dgst[0];
tmps[gid].dgst[i + 1] = dgst[1];
tmps[gid].dgst[i + 2] = dgst[2];
tmps[gid].dgst[i + 3] = dgst[3];
tmps[gid].dgst[i + 4] = dgst[4];
tmps[gid].dgst[i + 5] = dgst[5];
tmps[gid].dgst[i + 6] = dgst[6];
tmps[gid].dgst[i + 7] = dgst[7];
tmps[gid].out[i + 0] = dgst[0];
tmps[gid].out[i + 1] = dgst[1];
tmps[gid].out[i + 2] = dgst[2];
tmps[gid].out[i + 3] = dgst[3];
tmps[gid].out[i + 4] = dgst[4];
tmps[gid].out[i + 5] = dgst[5];
tmps[gid].out[i + 6] = dgst[6];
tmps[gid].out[i + 7] = dgst[7];
}
}
__kernel void m06223_loop (__global pw_t *pws, __global kernel_rule_t *rules_buf, __global comb_t *combs_buf, __global bf_t *bfs_buf, __global tc64_tmp_t *tmps, __global void *hooks, __global u32 *bitmaps_buf_s1_a, __global u32 *bitmaps_buf_s1_b, __global u32 *bitmaps_buf_s1_c, __global u32 *bitmaps_buf_s1_d, __global u32 *bitmaps_buf_s2_a, __global u32 *bitmaps_buf_s2_b, __global u32 *bitmaps_buf_s2_c, __global u32 *bitmaps_buf_s2_d, __global plain_t *plains_buf, __global digest_t *digests_buf, __global u32 *hashes_shown, __global salt_t *salt_bufs, __global tc_t *esalt_bufs, __global u32 *d_return_buf, __global u32 *d_scryptV_buf, const u32 bitmap_mask, const u32 bitmap_shift1, const u32 bitmap_shift2, const u32 salt_pos, const u32 loop_pos, const u32 loop_cnt, const u32 il_cnt, const u32 digests_cnt, const u32 digests_offset, const u32 combs_mode, const u32 gid_max)
{
const u32 truecrypt_mdlen = salt_bufs[0].truecrypt_mdlen;
const u32 gid = get_global_id (0);
if (gid >= gid_max) return;
u64 ipad[8];
ipad[0] = tmps[gid].ipad[0];
ipad[1] = tmps[gid].ipad[1];
ipad[2] = tmps[gid].ipad[2];
ipad[3] = tmps[gid].ipad[3];
ipad[4] = tmps[gid].ipad[4];
ipad[5] = tmps[gid].ipad[5];
ipad[6] = tmps[gid].ipad[6];
ipad[7] = tmps[gid].ipad[7];
u64 opad[8];
opad[0] = tmps[gid].opad[0];
opad[1] = tmps[gid].opad[1];
opad[2] = tmps[gid].opad[2];
opad[3] = tmps[gid].opad[3];
opad[4] = tmps[gid].opad[4];
opad[5] = tmps[gid].opad[5];
opad[6] = tmps[gid].opad[6];
opad[7] = tmps[gid].opad[7];
for (u32 i = 0; i < (truecrypt_mdlen / 8 / 8); i += 8)
{
u64 dgst[8];
dgst[0] = tmps[gid].dgst[i + 0];
dgst[1] = tmps[gid].dgst[i + 1];
dgst[2] = tmps[gid].dgst[i + 2];
dgst[3] = tmps[gid].dgst[i + 3];
dgst[4] = tmps[gid].dgst[i + 4];
dgst[5] = tmps[gid].dgst[i + 5];
dgst[6] = tmps[gid].dgst[i + 6];
dgst[7] = tmps[gid].dgst[i + 7];
u64 out[8];
out[0] = tmps[gid].out[i + 0];
out[1] = tmps[gid].out[i + 1];
out[2] = tmps[gid].out[i + 2];
out[3] = tmps[gid].out[i + 3];
out[4] = tmps[gid].out[i + 4];
out[5] = tmps[gid].out[i + 5];
out[6] = tmps[gid].out[i + 6];
out[7] = tmps[gid].out[i + 7];
for (u32 j = 0; j < loop_cnt; j++)
{
u64 w[16];
w[ 0] = dgst[0];
w[ 1] = dgst[1];
w[ 2] = dgst[2];
w[ 3] = dgst[3];
w[ 4] = dgst[4];
w[ 5] = dgst[5];
w[ 6] = dgst[6];
w[ 7] = dgst[7];
w[ 8] = 0x8000000000000000;
w[ 9] = 0;
w[10] = 0;
w[11] = 0;
w[12] = 0;
w[13] = 0;
w[14] = 0;
w[15] = (128 + 64) * 8;
hmac_run (w, ipad, opad, dgst);
out[0] ^= dgst[0];
out[1] ^= dgst[1];
out[2] ^= dgst[2];
out[3] ^= dgst[3];
out[4] ^= dgst[4];
out[5] ^= dgst[5];
out[6] ^= dgst[6];
out[7] ^= dgst[7];
}
tmps[gid].dgst[i + 0] = dgst[0];
tmps[gid].dgst[i + 1] = dgst[1];
tmps[gid].dgst[i + 2] = dgst[2];
tmps[gid].dgst[i + 3] = dgst[3];
tmps[gid].dgst[i + 4] = dgst[4];
tmps[gid].dgst[i + 5] = dgst[5];
tmps[gid].dgst[i + 6] = dgst[6];
tmps[gid].dgst[i + 7] = dgst[7];
tmps[gid].out[i + 0] = out[0];
tmps[gid].out[i + 1] = out[1];
tmps[gid].out[i + 2] = out[2];
tmps[gid].out[i + 3] = out[3];
tmps[gid].out[i + 4] = out[4];
tmps[gid].out[i + 5] = out[5];
tmps[gid].out[i + 6] = out[6];
tmps[gid].out[i + 7] = out[7];
}
}
__kernel void m06223_comp (__global pw_t *pws, __global kernel_rule_t *rules_buf, __global comb_t *combs_buf, __global bf_t *bfs_buf, __global tc64_tmp_t *tmps, __global void *hooks, __global u32 *bitmaps_buf_s1_a, __global u32 *bitmaps_buf_s1_b, __global u32 *bitmaps_buf_s1_c, __global u32 *bitmaps_buf_s1_d, __global u32 *bitmaps_buf_s2_a, __global u32 *bitmaps_buf_s2_b, __global u32 *bitmaps_buf_s2_c, __global u32 *bitmaps_buf_s2_d, __global plain_t *plains_buf, __global digest_t *digests_buf, __global u32 *hashes_shown, __global salt_t *salt_bufs, __global tc_t *esalt_bufs, __global u32 *d_return_buf, __global u32 *d_scryptV_buf, const u32 bitmap_mask, const u32 bitmap_shift1, const u32 bitmap_shift2, const u32 salt_pos, const u32 loop_pos, const u32 loop_cnt, const u32 il_cnt, const u32 digests_cnt, const u32 digests_offset, const u32 combs_mode, const u32 gid_max)
{
/**
* base
*/
const u32 gid = get_global_id (0);
const u32 lid = get_local_id (0);
if (gid >= gid_max) return;
u32 ukey1[8];
ukey1[0] = swap32 (h32_from_64 (tmps[gid].out[ 0]));
ukey1[1] = swap32 (l32_from_64 (tmps[gid].out[ 0]));
ukey1[2] = swap32 (h32_from_64 (tmps[gid].out[ 1]));
ukey1[3] = swap32 (l32_from_64 (tmps[gid].out[ 1]));
ukey1[4] = swap32 (h32_from_64 (tmps[gid].out[ 2]));
ukey1[5] = swap32 (l32_from_64 (tmps[gid].out[ 2]));
ukey1[6] = swap32 (h32_from_64 (tmps[gid].out[ 3]));
ukey1[7] = swap32 (l32_from_64 (tmps[gid].out[ 3]));
u32 ukey2[8];
ukey2[0] = swap32 (h32_from_64 (tmps[gid].out[ 4]));
ukey2[1] = swap32 (l32_from_64 (tmps[gid].out[ 4]));
ukey2[2] = swap32 (h32_from_64 (tmps[gid].out[ 5]));
ukey2[3] = swap32 (l32_from_64 (tmps[gid].out[ 5]));
ukey2[4] = swap32 (h32_from_64 (tmps[gid].out[ 6]));
ukey2[5] = swap32 (l32_from_64 (tmps[gid].out[ 6]));
ukey2[6] = swap32 (h32_from_64 (tmps[gid].out[ 7]));
ukey2[7] = swap32 (l32_from_64 (tmps[gid].out[ 7]));
u32 data[4];
data[0] = esalt_bufs[0].data_buf[0];
data[1] = esalt_bufs[0].data_buf[1];
data[2] = esalt_bufs[0].data_buf[2];
data[3] = esalt_bufs[0].data_buf[3];
u32 tmp[4];
{
tmp[0] = data[0];
tmp[1] = data[1];
tmp[2] = data[2];
tmp[3] = data[3];
aes256_decrypt_xts (ukey1, ukey2, tmp, tmp);
if (((tmp[0] == 0x45555254) && (tmp[3] == 0)) || ((tmp[0] == 0x45555254) && ((tmp[1] >> 16) <= 5)))
{
mark_hash (plains_buf, hashes_shown, 0, gid, 0);
d_return_buf[lid] = 1;
}
}
{
tmp[0] = data[0];
tmp[1] = data[1];
tmp[2] = data[2];
tmp[3] = data[3];
serpent256_decrypt_xts (ukey1, ukey2, tmp, tmp);
if (((tmp[0] == 0x45555254) && (tmp[3] == 0)) || ((tmp[0] == 0x45555254) && ((tmp[1] >> 16) <= 5)))
{
mark_hash (plains_buf, hashes_shown, 0, gid, 0);
d_return_buf[lid] = 1;
}
}
{
tmp[0] = data[0];
tmp[1] = data[1];
tmp[2] = data[2];
tmp[3] = data[3];
twofish256_decrypt_xts (ukey1, ukey2, tmp, tmp);
if (((tmp[0] == 0x45555254) && (tmp[3] == 0)) || ((tmp[0] == 0x45555254) && ((tmp[1] >> 16) <= 5)))
{
mark_hash (plains_buf, hashes_shown, 0, gid, 0);
d_return_buf[lid] = 1;
}
}
u32 ukey3[8];
ukey3[0] = swap32 (h32_from_64 (tmps[gid].out[ 8]));
ukey3[1] = swap32 (l32_from_64 (tmps[gid].out[ 8]));
ukey3[2] = swap32 (h32_from_64 (tmps[gid].out[ 9]));
ukey3[3] = swap32 (l32_from_64 (tmps[gid].out[ 9]));
ukey3[4] = swap32 (h32_from_64 (tmps[gid].out[10]));
ukey3[5] = swap32 (l32_from_64 (tmps[gid].out[10]));
ukey3[6] = swap32 (h32_from_64 (tmps[gid].out[11]));
ukey3[7] = swap32 (l32_from_64 (tmps[gid].out[11]));
u32 ukey4[8];
ukey4[0] = swap32 (h32_from_64 (tmps[gid].out[12]));
ukey4[1] = swap32 (l32_from_64 (tmps[gid].out[12]));
ukey4[2] = swap32 (h32_from_64 (tmps[gid].out[13]));
ukey4[3] = swap32 (l32_from_64 (tmps[gid].out[13]));
ukey4[4] = swap32 (h32_from_64 (tmps[gid].out[14]));
ukey4[5] = swap32 (l32_from_64 (tmps[gid].out[14]));
ukey4[6] = swap32 (h32_from_64 (tmps[gid].out[15]));
ukey4[7] = swap32 (l32_from_64 (tmps[gid].out[15]));
{
tmp[0] = data[0];
tmp[1] = data[1];
tmp[2] = data[2];
tmp[3] = data[3];
aes256_decrypt_xts (ukey2, ukey4, tmp, tmp);
twofish256_decrypt_xts (ukey1, ukey3, tmp, tmp);
if (((tmp[0] == 0x45555254) && (tmp[3] == 0)) || ((tmp[0] == 0x45555254) && ((tmp[1] >> 16) <= 5)))
{
mark_hash (plains_buf, hashes_shown, 0, gid, 0);
d_return_buf[lid] = 1;
}
}
{
tmp[0] = data[0];
tmp[1] = data[1];
tmp[2] = data[2];
tmp[3] = data[3];
serpent256_decrypt_xts (ukey2, ukey4, tmp, tmp);
aes256_decrypt_xts (ukey1, ukey3, tmp, tmp);
if (((tmp[0] == 0x45555254) && (tmp[3] == 0)) || ((tmp[0] == 0x45555254) && ((tmp[1] >> 16) <= 5)))
{
mark_hash (plains_buf, hashes_shown, 0, gid, 0);
d_return_buf[lid] = 1;
}
}
{
tmp[0] = data[0];
tmp[1] = data[1];
tmp[2] = data[2];
tmp[3] = data[3];
twofish256_decrypt_xts (ukey2, ukey4, tmp, tmp);
serpent256_decrypt_xts (ukey1, ukey3, tmp, tmp);
if (((tmp[0] == 0x45555254) && (tmp[3] == 0)) || ((tmp[0] == 0x45555254) && ((tmp[1] >> 16) <= 5)))
{
mark_hash (plains_buf, hashes_shown, 0, gid, 0);
d_return_buf[lid] = 1;
}
}
u32 ukey5[8];
ukey5[0] = swap32 (h32_from_64 (tmps[gid].out[16]));
ukey5[1] = swap32 (l32_from_64 (tmps[gid].out[16]));
ukey5[2] = swap32 (h32_from_64 (tmps[gid].out[17]));
ukey5[3] = swap32 (l32_from_64 (tmps[gid].out[17]));
ukey5[4] = swap32 (h32_from_64 (tmps[gid].out[18]));
ukey5[5] = swap32 (l32_from_64 (tmps[gid].out[18]));
ukey5[6] = swap32 (h32_from_64 (tmps[gid].out[19]));
ukey5[7] = swap32 (l32_from_64 (tmps[gid].out[19]));
u32 ukey6[8];
ukey6[0] = swap32 (h32_from_64 (tmps[gid].out[20]));
ukey6[1] = swap32 (l32_from_64 (tmps[gid].out[20]));
ukey6[2] = swap32 (h32_from_64 (tmps[gid].out[21]));
ukey6[3] = swap32 (l32_from_64 (tmps[gid].out[21]));
ukey6[4] = swap32 (h32_from_64 (tmps[gid].out[22]));
ukey6[5] = swap32 (l32_from_64 (tmps[gid].out[22]));
ukey6[6] = swap32 (h32_from_64 (tmps[gid].out[23]));
ukey6[7] = swap32 (l32_from_64 (tmps[gid].out[23]));
{
tmp[0] = data[0];
tmp[1] = data[1];
tmp[2] = data[2];
tmp[3] = data[3];
aes256_decrypt_xts (ukey3, ukey6, tmp, tmp);
twofish256_decrypt_xts (ukey2, ukey5, tmp, tmp);
serpent256_decrypt_xts (ukey1, ukey4, tmp, tmp);
if (((tmp[0] == 0x45555254) && (tmp[3] == 0)) || ((tmp[0] == 0x45555254) && ((tmp[1] >> 16) <= 5)))
{
mark_hash (plains_buf, hashes_shown, 0, gid, 0);
d_return_buf[lid] = 1;
}
}
{
tmp[0] = data[0];
tmp[1] = data[1];
tmp[2] = data[2];
tmp[3] = data[3];
serpent256_decrypt_xts (ukey3, ukey6, tmp, tmp);
twofish256_decrypt_xts (ukey2, ukey5, tmp, tmp);
aes256_decrypt_xts (ukey1, ukey4, tmp, tmp);
if (((tmp[0] == 0x45555254) && (tmp[3] == 0)) || ((tmp[0] == 0x45555254) && ((tmp[1] >> 16) <= 5)))
{
mark_hash (plains_buf, hashes_shown, 0, gid, 0);
d_return_buf[lid] = 1;
}
}
}