mirror of
https://github.com/hashcat/hashcat.git
synced 2024-11-18 14:18:11 +00:00
3979 lines
114 KiB
Common Lisp
3979 lines
114 KiB
Common Lisp
|
|
// cbc-essiv
|
|
|
|
static void aes128_decrypt_cbc (const u32 *ks1, const u32 *in, u32 *out, u32 *essiv, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
aes128_decrypt (ks1, in, out, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
out[0] ^= essiv[0];
|
|
out[1] ^= essiv[1];
|
|
out[2] ^= essiv[2];
|
|
out[3] ^= essiv[3];
|
|
|
|
essiv[0] = in[0];
|
|
essiv[1] = in[1];
|
|
essiv[2] = in[2];
|
|
essiv[3] = in[3];
|
|
}
|
|
|
|
static void aes256_decrypt_cbc (const u32 *ks1, const u32 *in, u32 *out, u32 *essiv, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
aes256_decrypt (ks1, in, out, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
out[0] ^= essiv[0];
|
|
out[1] ^= essiv[1];
|
|
out[2] ^= essiv[2];
|
|
out[3] ^= essiv[3];
|
|
|
|
essiv[0] = in[0];
|
|
essiv[1] = in[1];
|
|
essiv[2] = in[2];
|
|
essiv[3] = in[3];
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv128 (__global u32 *in, u32 *out, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
int idx_out = 0;
|
|
|
|
for (int i = 0; i < 32; i++)
|
|
{
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
out[idx_out++] = data_out[0];
|
|
out[idx_out++] = data_out[1];
|
|
out[idx_out++] = data_out[2];
|
|
out[idx_out++] = data_out[3];
|
|
}
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv128_mk_sha1 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 32; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha1_diffuse16 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv128_mk_sha1_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 32 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha1_diffuse16 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha1_diffuse16()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv128_mk_sha256 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 32; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha256_diffuse16 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv128_mk_sha256_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 32 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha256_diffuse16 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha256_diffuse16()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv128_mk_sha512 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 32; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha512_diffuse16 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv128_mk_sha512_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 32 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha512_diffuse16 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha512_diffuse16()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv128_mk_ripemd160 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 32; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_ripemd160_diffuse16 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv128_mk_ripemd160_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 32 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_ripemd160_diffuse16 (mk);
|
|
}
|
|
|
|
// this one has no AF_ripemd160_diffuse16()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv256 (__global u32 *in, u32 *out, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
int idx_out = 0;
|
|
|
|
for (int i = 0; i < 16; i++)
|
|
{
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
out[idx_out++] = data_out[0];
|
|
out[idx_out++] = data_out[1];
|
|
out[idx_out++] = data_out[2];
|
|
out[idx_out++] = data_out[3];
|
|
}
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv256_mk_sha1 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha1_diffuse32 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv256_mk_sha1_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha1_diffuse32 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha1_diffuse32()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv256_mk_sha256 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha256_diffuse32 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv256_mk_sha256_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha256_diffuse32 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha256_diffuse32()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv256_mk_sha512 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha512_diffuse32 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv256_mk_sha512_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha512_diffuse32 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha512_diffuse32()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv256_mk_ripemd160 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_ripemd160_diffuse32 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_essiv256_mk_ripemd160_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 essiv[4];
|
|
|
|
aes256_encrypt (ks2, S, essiv, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_ripemd160_diffuse32 (mk);
|
|
}
|
|
|
|
// this one has no AF_ripemd160_diffuse32()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, essiv, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
// cbc-plain
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain128 (__global u32 *in, u32 *out, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
int idx_out = 0;
|
|
|
|
for (int i = 0; i < 32; i++)
|
|
{
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
out[idx_out++] = data_out[0];
|
|
out[idx_out++] = data_out[1];
|
|
out[idx_out++] = data_out[2];
|
|
out[idx_out++] = data_out[3];
|
|
}
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain128_mk_sha1 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 32; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha1_diffuse16 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain128_mk_sha1_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 32 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha1_diffuse16 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha1_diffuse16()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain128_mk_sha256 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 32; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha256_diffuse16 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain128_mk_sha256_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 32 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha256_diffuse16 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha256_diffuse16()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain128_mk_sha512 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 32; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha512_diffuse16 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain128_mk_sha512_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 32 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha512_diffuse16 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha512_diffuse16()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain128_mk_ripemd160 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 32; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_ripemd160_diffuse16 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain128_mk_ripemd160_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 32 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_ripemd160_diffuse16 (mk);
|
|
}
|
|
|
|
// this one has no AF_ripemd160_diffuse16()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 1; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain256 (__global u32 *in, u32 *out, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
int idx_out = 0;
|
|
|
|
for (int i = 0; i < 16; i++)
|
|
{
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
out[idx_out++] = data_out[0];
|
|
out[idx_out++] = data_out[1];
|
|
out[idx_out++] = data_out[2];
|
|
out[idx_out++] = data_out[3];
|
|
}
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain256_mk_sha1 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha1_diffuse32 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain256_mk_sha1_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha1_diffuse32 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha1_diffuse32()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain256_mk_sha256 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha256_diffuse32 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain256_mk_sha256_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha256_diffuse32 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha256_diffuse32()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain256_mk_sha512 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha512_diffuse32 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain256_mk_sha512_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha512_diffuse32 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha512_diffuse32()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain256_mk_ripemd160 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_ripemd160_diffuse32 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_cbc_plain256_mk_ripemd160_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_ripemd160_diffuse32 (mk);
|
|
}
|
|
|
|
// this one has no AF_ripemd160_diffuse32()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_cbc (ks1, data_in, data_out, S, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
// xts-plain
|
|
|
|
static void aes128_decrypt_xts (const u32 *ks1, const u32 *in, u32 *out, u32 *T, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
out[0] = in[0];
|
|
out[1] = in[1];
|
|
out[2] = in[2];
|
|
out[3] = in[3];
|
|
|
|
out[0] ^= T[0];
|
|
out[1] ^= T[1];
|
|
out[2] ^= T[2];
|
|
out[3] ^= T[3];
|
|
|
|
aes128_decrypt (ks1, out, out, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
out[0] ^= T[0];
|
|
out[1] ^= T[1];
|
|
out[2] ^= T[2];
|
|
out[3] ^= T[3];
|
|
|
|
xts_mul2 (T, T);
|
|
}
|
|
|
|
static void aes256_decrypt_xts (const u32 *ks1, const u32 *in, u32 *out, u32 *T, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
out[0] = in[0];
|
|
out[1] = in[1];
|
|
out[2] = in[2];
|
|
out[3] = in[3];
|
|
|
|
out[0] ^= T[0];
|
|
out[1] ^= T[1];
|
|
out[2] ^= T[2];
|
|
out[3] ^= T[3];
|
|
|
|
aes256_decrypt (ks1, out, out, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
out[0] ^= T[0];
|
|
out[1] ^= T[1];
|
|
out[2] ^= T[2];
|
|
out[3] ^= T[3];
|
|
|
|
xts_mul2 (T, T);
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain256 (__global u32 *in, u32 *out, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes128_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
int idx_out = 0;
|
|
|
|
for (int i = 0; i < 16; i++)
|
|
{
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
out[idx_out++] = data_out[0];
|
|
out[idx_out++] = data_out[1];
|
|
out[idx_out++] = data_out[2];
|
|
out[idx_out++] = data_out[3];
|
|
}
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain256_mk_sha1 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes128_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha1_diffuse32 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain256_mk_sha1_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes128_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha1_diffuse32 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha1_diffuse32()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain256_mk_sha256 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes128_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha256_diffuse32 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain256_mk_sha256_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes128_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha256_diffuse32 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha256_diffuse32()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain256_mk_sha512 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes128_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha512_diffuse32 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain256_mk_sha512_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes128_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha512_diffuse32 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha512_diffuse32()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain256_mk_ripemd160 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes128_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_ripemd160_diffuse32 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain256_mk_ripemd160_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes128_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 16 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_ripemd160_diffuse32 (mk);
|
|
}
|
|
|
|
// this one has no AF_ripemd160_diffuse32()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 2; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes128_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain512 (__global u32 *in, u32 *out, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes256_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
int idx_out = 0;
|
|
|
|
for (int i = 0; i < 8; i++)
|
|
{
|
|
for (int block = 0; block < 4; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
out[idx_out++] = data_out[0];
|
|
out[idx_out++] = data_out[1];
|
|
out[idx_out++] = data_out[2];
|
|
out[idx_out++] = data_out[3];
|
|
}
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain512_mk_sha1 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes256_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 8; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 4; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha1_diffuse64 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain512_mk_sha1_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes256_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 8 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 4; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha1_diffuse64 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha1_diffuse64()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 4; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain512_mk_sha256 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes256_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 8; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 4; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha256_diffuse64 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain512_mk_sha256_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes256_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 8 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 4; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha256_diffuse64 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha256_diffuse64()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 4; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain512_mk_sha512 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes256_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 8; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 4; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha512_diffuse64 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain512_mk_sha512_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes256_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 8 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 4; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_sha512_diffuse64 (mk);
|
|
}
|
|
|
|
// this one has no AF_sha512_diffuse64()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 4; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain512_mk_ripemd160 (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes256_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 8; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 4; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_ripemd160_diffuse64 (mk);
|
|
}
|
|
}
|
|
|
|
static void luks_decrypt_sector_aes_xts_plain512_mk_ripemd160_final (__global u32 *in, u32 *mk, const u32 *ks1, const u32 *ks2, const u32 sector, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
u32 S[4] = { sector, 0, 0, 0 };
|
|
|
|
u32 T[4];
|
|
|
|
aes256_encrypt (ks2, S, T, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int idx_in = 0;
|
|
|
|
for (int i = 0; i < 8 - 1; i++)
|
|
{
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 4; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
|
|
AF_ripemd160_diffuse64 (mk);
|
|
}
|
|
|
|
// this one has no AF_ripemd160_diffuse64()
|
|
|
|
int idx_mk = 0;
|
|
|
|
for (int block = 0; block < 4; block++)
|
|
{
|
|
u32 data_in[4];
|
|
|
|
data_in[0] = in[idx_in++];
|
|
data_in[1] = in[idx_in++];
|
|
data_in[2] = in[idx_in++];
|
|
data_in[3] = in[idx_in++];
|
|
|
|
u32 data_out[4];
|
|
|
|
aes256_decrypt_xts (ks1, data_in, data_out, T, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
mk[idx_mk++] ^= data_out[0];
|
|
mk[idx_mk++] ^= data_out[1];
|
|
mk[idx_mk++] ^= data_out[2];
|
|
mk[idx_mk++] ^= data_out[3];
|
|
}
|
|
}
|
|
|
|
// luks helper
|
|
|
|
static void luks_af_sha1_then_aes_decrypt (__global luks_t *luks_bufs, __global luks_tmp_t *tmps, u32 *pt_buf, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
const u32 key_size = luks_bufs->key_size;
|
|
const u32 cipher_mode = luks_bufs->cipher_mode;
|
|
|
|
#define BITS_PER_AF (key_size * LUKS_STRIPES)
|
|
#define BITS_PER_SECTOR (512 * 8)
|
|
#define SECTOR_PER_AF (BITS_PER_AF / BITS_PER_SECTOR)
|
|
#define BLOCKS_PER_SECTOR (512 / 16)
|
|
#define OFFSET_PER_BLOCK (16 / 4)
|
|
#define OFFSET_PER_SECTOR (BLOCKS_PER_SECTOR * OFFSET_PER_BLOCK)
|
|
|
|
// decrypt AF data and do the AF merge inline
|
|
|
|
u32 mk[16] = { 0 };
|
|
|
|
if (cipher_mode == HC_LUKS_CIPHER_MODE_CBC_ESSIV)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_128)
|
|
{
|
|
u32 ukey[4];
|
|
|
|
ukey[0] = swap32_S (tmps->out32[0]);
|
|
ukey[1] = swap32_S (tmps->out32[1]);
|
|
ukey[2] = swap32_S (tmps->out32[2]);
|
|
ukey[3] = swap32_S (tmps->out32[3]);
|
|
|
|
u32 essivhash[8];
|
|
|
|
ESSIV_sha256_init128 (ukey, essivhash);
|
|
|
|
u32 ks1[44];
|
|
u32 ks2[60];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, essivhash, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_cbc_essiv128_mk_sha1 (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_cbc_essiv128_mk_sha1_final (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey[8];
|
|
|
|
ukey[0] = swap32_S (tmps->out32[0]);
|
|
ukey[1] = swap32_S (tmps->out32[1]);
|
|
ukey[2] = swap32_S (tmps->out32[2]);
|
|
ukey[3] = swap32_S (tmps->out32[3]);
|
|
ukey[4] = swap32_S (tmps->out32[4]);
|
|
ukey[5] = swap32_S (tmps->out32[5]);
|
|
ukey[6] = swap32_S (tmps->out32[6]);
|
|
ukey[7] = swap32_S (tmps->out32[7]);
|
|
|
|
u32 essivhash[8];
|
|
|
|
ESSIV_sha256_init256 (ukey, essivhash);
|
|
|
|
u32 ks1[60];
|
|
u32 ks2[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, essivhash, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_cbc_essiv256_mk_sha1 (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_cbc_essiv256_mk_sha1_final (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
else if (cipher_mode == HC_LUKS_CIPHER_MODE_CBC_PLAIN)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_128)
|
|
{
|
|
u32 ukey[4];
|
|
|
|
ukey[0] = swap32_S (tmps->out32[0]);
|
|
ukey[1] = swap32_S (tmps->out32[1]);
|
|
ukey[2] = swap32_S (tmps->out32[2]);
|
|
ukey[3] = swap32_S (tmps->out32[3]);
|
|
|
|
u32 ks1[44];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_cbc_plain128_mk_sha1 (luks_bufs->af_src_buf + offset, mk, ks1, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_cbc_plain128_mk_sha1_final (luks_bufs->af_src_buf + offset, mk, ks1, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey[8];
|
|
|
|
ukey[0] = swap32_S (tmps->out32[0]);
|
|
ukey[1] = swap32_S (tmps->out32[1]);
|
|
ukey[2] = swap32_S (tmps->out32[2]);
|
|
ukey[3] = swap32_S (tmps->out32[3]);
|
|
ukey[4] = swap32_S (tmps->out32[4]);
|
|
ukey[5] = swap32_S (tmps->out32[5]);
|
|
ukey[6] = swap32_S (tmps->out32[6]);
|
|
ukey[7] = swap32_S (tmps->out32[7]);
|
|
|
|
u32 ks1[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_cbc_plain256_mk_sha1 (luks_bufs->af_src_buf + offset, mk, ks1, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_cbc_plain256_mk_sha1_final (luks_bufs->af_src_buf + offset, mk, ks1, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
else if (cipher_mode == HC_LUKS_CIPHER_MODE_XTS_PLAIN)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey1[4];
|
|
|
|
ukey1[0] = swap32_S (tmps->out32[0]);
|
|
ukey1[1] = swap32_S (tmps->out32[1]);
|
|
ukey1[2] = swap32_S (tmps->out32[2]);
|
|
ukey1[3] = swap32_S (tmps->out32[3]);
|
|
|
|
u32 ukey2[4];
|
|
|
|
ukey2[0] = swap32_S (tmps->out32[4]);
|
|
ukey2[1] = swap32_S (tmps->out32[5]);
|
|
ukey2[2] = swap32_S (tmps->out32[6]);
|
|
ukey2[3] = swap32_S (tmps->out32[7]);
|
|
|
|
u32 ks1[44];
|
|
u32 ks2[44];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey1, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes128_set_encrypt_key (ks2, ukey2, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_xts_plain256_mk_sha1 (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_xts_plain256_mk_sha1_final (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_512)
|
|
{
|
|
u32 ukey1[8];
|
|
|
|
ukey1[0] = swap32_S (tmps->out32[ 0]);
|
|
ukey1[1] = swap32_S (tmps->out32[ 1]);
|
|
ukey1[2] = swap32_S (tmps->out32[ 2]);
|
|
ukey1[3] = swap32_S (tmps->out32[ 3]);
|
|
ukey1[4] = swap32_S (tmps->out32[ 4]);
|
|
ukey1[5] = swap32_S (tmps->out32[ 5]);
|
|
ukey1[6] = swap32_S (tmps->out32[ 6]);
|
|
ukey1[7] = swap32_S (tmps->out32[ 7]);
|
|
|
|
u32 ukey2[8];
|
|
|
|
ukey2[0] = swap32_S (tmps->out32[ 8]);
|
|
ukey2[1] = swap32_S (tmps->out32[ 9]);
|
|
ukey2[2] = swap32_S (tmps->out32[10]);
|
|
ukey2[3] = swap32_S (tmps->out32[11]);
|
|
ukey2[4] = swap32_S (tmps->out32[12]);
|
|
ukey2[5] = swap32_S (tmps->out32[13]);
|
|
ukey2[6] = swap32_S (tmps->out32[14]);
|
|
ukey2[7] = swap32_S (tmps->out32[15]);
|
|
|
|
u32 ks1[60];
|
|
u32 ks2[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey1, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, ukey2, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_xts_plain512_mk_sha1 (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_xts_plain512_mk_sha1_final (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
|
|
// decrypt payload data
|
|
|
|
if (cipher_mode == HC_LUKS_CIPHER_MODE_CBC_ESSIV)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_128)
|
|
{
|
|
u32 ukey[4];
|
|
|
|
ukey[0] = mk[0];
|
|
ukey[1] = mk[1];
|
|
ukey[2] = mk[2];
|
|
ukey[3] = mk[3];
|
|
|
|
u32 essivhash[8];
|
|
|
|
ESSIV_sha256_init128 (ukey, essivhash);
|
|
|
|
u32 ks1[44];
|
|
u32 ks2[60];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, essivhash, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
luks_decrypt_sector_aes_cbc_essiv128 (luks_bufs->ct_buf, pt_buf, ks1, ks2, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey[8];
|
|
|
|
ukey[0] = mk[0];
|
|
ukey[1] = mk[1];
|
|
ukey[2] = mk[2];
|
|
ukey[3] = mk[3];
|
|
ukey[4] = mk[4];
|
|
ukey[5] = mk[5];
|
|
ukey[6] = mk[6];
|
|
ukey[7] = mk[7];
|
|
|
|
u32 essivhash[8];
|
|
|
|
ESSIV_sha256_init256 (ukey, essivhash);
|
|
|
|
u32 ks1[60];
|
|
u32 ks2[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, essivhash, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
luks_decrypt_sector_aes_cbc_essiv256 (luks_bufs->ct_buf, pt_buf, ks1, ks2, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
else if (cipher_mode == HC_LUKS_CIPHER_MODE_CBC_PLAIN)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_128)
|
|
{
|
|
u32 ukey[4];
|
|
|
|
ukey[0] = mk[0];
|
|
ukey[1] = mk[1];
|
|
ukey[2] = mk[2];
|
|
ukey[3] = mk[3];
|
|
|
|
u32 ks1[44];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
luks_decrypt_sector_aes_cbc_plain128 (luks_bufs->ct_buf, pt_buf, ks1, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey[8];
|
|
|
|
ukey[0] = mk[0];
|
|
ukey[1] = mk[1];
|
|
ukey[2] = mk[2];
|
|
ukey[3] = mk[3];
|
|
ukey[4] = mk[4];
|
|
ukey[5] = mk[5];
|
|
ukey[6] = mk[6];
|
|
ukey[7] = mk[7];
|
|
|
|
u32 ks1[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
luks_decrypt_sector_aes_cbc_plain256 (luks_bufs->ct_buf, pt_buf, ks1, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
else if (cipher_mode == HC_LUKS_CIPHER_MODE_XTS_PLAIN)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey1[4];
|
|
|
|
ukey1[0] = mk[0];
|
|
ukey1[1] = mk[1];
|
|
ukey1[2] = mk[2];
|
|
ukey1[3] = mk[3];
|
|
|
|
u32 ukey2[4];
|
|
|
|
ukey2[0] = mk[4];
|
|
ukey2[1] = mk[5];
|
|
ukey2[2] = mk[6];
|
|
ukey2[3] = mk[7];
|
|
|
|
u32 ks1[44];
|
|
u32 ks2[44];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey1, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes128_set_encrypt_key (ks2, ukey2, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
luks_decrypt_sector_aes_xts_plain256 (luks_bufs->ct_buf, pt_buf, ks1, ks2, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_512)
|
|
{
|
|
u32 ukey1[8];
|
|
|
|
ukey1[0] = mk[ 0];
|
|
ukey1[1] = mk[ 1];
|
|
ukey1[2] = mk[ 2];
|
|
ukey1[3] = mk[ 3];
|
|
ukey1[4] = mk[ 4];
|
|
ukey1[5] = mk[ 5];
|
|
ukey1[6] = mk[ 6];
|
|
ukey1[7] = mk[ 7];
|
|
|
|
u32 ukey2[8];
|
|
|
|
ukey2[0] = mk[ 8];
|
|
ukey2[1] = mk[ 9];
|
|
ukey2[2] = mk[10];
|
|
ukey2[3] = mk[11];
|
|
ukey2[4] = mk[12];
|
|
ukey2[5] = mk[13];
|
|
ukey2[6] = mk[14];
|
|
ukey2[7] = mk[15];
|
|
|
|
u32 ks1[60];
|
|
u32 ks2[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey1, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, ukey2, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
luks_decrypt_sector_aes_xts_plain512 (luks_bufs->ct_buf, pt_buf, ks1, ks2, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void luks_af_sha256_then_aes_decrypt (__global luks_t *luks_bufs, __global luks_tmp_t *tmps, u32 *pt_buf, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
const u32 key_size = luks_bufs->key_size;
|
|
const u32 cipher_mode = luks_bufs->cipher_mode;
|
|
|
|
#define BITS_PER_AF (key_size * LUKS_STRIPES)
|
|
#define BITS_PER_SECTOR (512 * 8)
|
|
#define SECTOR_PER_AF (BITS_PER_AF / BITS_PER_SECTOR)
|
|
#define BLOCKS_PER_SECTOR (512 / 16)
|
|
#define OFFSET_PER_BLOCK (16 / 4)
|
|
#define OFFSET_PER_SECTOR (BLOCKS_PER_SECTOR * OFFSET_PER_BLOCK)
|
|
|
|
// decrypt AF data and do the AF merge inline
|
|
|
|
u32 mk[16] = { 0 };
|
|
|
|
if (cipher_mode == HC_LUKS_CIPHER_MODE_CBC_ESSIV)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_128)
|
|
{
|
|
u32 ukey[4];
|
|
|
|
ukey[0] = swap32_S (tmps->out32[0]);
|
|
ukey[1] = swap32_S (tmps->out32[1]);
|
|
ukey[2] = swap32_S (tmps->out32[2]);
|
|
ukey[3] = swap32_S (tmps->out32[3]);
|
|
|
|
u32 essivhash[8];
|
|
|
|
ESSIV_sha256_init128 (ukey, essivhash);
|
|
|
|
u32 ks1[44];
|
|
u32 ks2[60];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, essivhash, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_cbc_essiv128_mk_sha256 (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_cbc_essiv128_mk_sha256_final (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey[8];
|
|
|
|
ukey[0] = swap32_S (tmps->out32[0]);
|
|
ukey[1] = swap32_S (tmps->out32[1]);
|
|
ukey[2] = swap32_S (tmps->out32[2]);
|
|
ukey[3] = swap32_S (tmps->out32[3]);
|
|
ukey[4] = swap32_S (tmps->out32[4]);
|
|
ukey[5] = swap32_S (tmps->out32[5]);
|
|
ukey[6] = swap32_S (tmps->out32[6]);
|
|
ukey[7] = swap32_S (tmps->out32[7]);
|
|
|
|
u32 essivhash[8];
|
|
|
|
ESSIV_sha256_init256 (ukey, essivhash);
|
|
|
|
u32 ks1[60];
|
|
u32 ks2[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, essivhash, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_cbc_essiv256_mk_sha256 (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_cbc_essiv256_mk_sha256_final (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
else if (cipher_mode == HC_LUKS_CIPHER_MODE_CBC_PLAIN)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_128)
|
|
{
|
|
u32 ukey[4];
|
|
|
|
ukey[0] = swap32_S (tmps->out32[0]);
|
|
ukey[1] = swap32_S (tmps->out32[1]);
|
|
ukey[2] = swap32_S (tmps->out32[2]);
|
|
ukey[3] = swap32_S (tmps->out32[3]);
|
|
|
|
u32 ks1[44];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_cbc_plain128_mk_sha256 (luks_bufs->af_src_buf + offset, mk, ks1, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_cbc_plain128_mk_sha256_final (luks_bufs->af_src_buf + offset, mk, ks1, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey[8];
|
|
|
|
ukey[0] = swap32_S (tmps->out32[0]);
|
|
ukey[1] = swap32_S (tmps->out32[1]);
|
|
ukey[2] = swap32_S (tmps->out32[2]);
|
|
ukey[3] = swap32_S (tmps->out32[3]);
|
|
ukey[4] = swap32_S (tmps->out32[4]);
|
|
ukey[5] = swap32_S (tmps->out32[5]);
|
|
ukey[6] = swap32_S (tmps->out32[6]);
|
|
ukey[7] = swap32_S (tmps->out32[7]);
|
|
|
|
u32 ks1[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_cbc_plain256_mk_sha256 (luks_bufs->af_src_buf + offset, mk, ks1, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_cbc_plain256_mk_sha256_final (luks_bufs->af_src_buf + offset, mk, ks1, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
else if (cipher_mode == HC_LUKS_CIPHER_MODE_XTS_PLAIN)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey1[4];
|
|
|
|
ukey1[0] = swap32_S (tmps->out32[0]);
|
|
ukey1[1] = swap32_S (tmps->out32[1]);
|
|
ukey1[2] = swap32_S (tmps->out32[2]);
|
|
ukey1[3] = swap32_S (tmps->out32[3]);
|
|
|
|
u32 ukey2[4];
|
|
|
|
ukey2[0] = swap32_S (tmps->out32[4]);
|
|
ukey2[1] = swap32_S (tmps->out32[5]);
|
|
ukey2[2] = swap32_S (tmps->out32[6]);
|
|
ukey2[3] = swap32_S (tmps->out32[7]);
|
|
|
|
u32 ks1[44];
|
|
u32 ks2[44];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey1, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes128_set_encrypt_key (ks2, ukey2, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_xts_plain256_mk_sha256 (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_xts_plain256_mk_sha256_final (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_512)
|
|
{
|
|
u32 ukey1[8];
|
|
|
|
ukey1[0] = swap32_S (tmps->out32[ 0]);
|
|
ukey1[1] = swap32_S (tmps->out32[ 1]);
|
|
ukey1[2] = swap32_S (tmps->out32[ 2]);
|
|
ukey1[3] = swap32_S (tmps->out32[ 3]);
|
|
ukey1[4] = swap32_S (tmps->out32[ 4]);
|
|
ukey1[5] = swap32_S (tmps->out32[ 5]);
|
|
ukey1[6] = swap32_S (tmps->out32[ 6]);
|
|
ukey1[7] = swap32_S (tmps->out32[ 7]);
|
|
|
|
u32 ukey2[8];
|
|
|
|
ukey2[0] = swap32_S (tmps->out32[ 8]);
|
|
ukey2[1] = swap32_S (tmps->out32[ 9]);
|
|
ukey2[2] = swap32_S (tmps->out32[10]);
|
|
ukey2[3] = swap32_S (tmps->out32[11]);
|
|
ukey2[4] = swap32_S (tmps->out32[12]);
|
|
ukey2[5] = swap32_S (tmps->out32[13]);
|
|
ukey2[6] = swap32_S (tmps->out32[14]);
|
|
ukey2[7] = swap32_S (tmps->out32[15]);
|
|
|
|
u32 ks1[60];
|
|
u32 ks2[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey1, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, ukey2, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_xts_plain512_mk_sha256 (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_xts_plain512_mk_sha256_final (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
|
|
// decrypt payload data
|
|
|
|
if (cipher_mode == HC_LUKS_CIPHER_MODE_CBC_ESSIV)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_128)
|
|
{
|
|
u32 ukey[4];
|
|
|
|
ukey[0] = mk[0];
|
|
ukey[1] = mk[1];
|
|
ukey[2] = mk[2];
|
|
ukey[3] = mk[3];
|
|
|
|
u32 essivhash[8];
|
|
|
|
ESSIV_sha256_init128 (ukey, essivhash);
|
|
|
|
u32 ks1[44];
|
|
u32 ks2[60];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, essivhash, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
luks_decrypt_sector_aes_cbc_essiv128 (luks_bufs->ct_buf, pt_buf, ks1, ks2, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey[8];
|
|
|
|
ukey[0] = mk[0];
|
|
ukey[1] = mk[1];
|
|
ukey[2] = mk[2];
|
|
ukey[3] = mk[3];
|
|
ukey[4] = mk[4];
|
|
ukey[5] = mk[5];
|
|
ukey[6] = mk[6];
|
|
ukey[7] = mk[7];
|
|
|
|
u32 essivhash[8];
|
|
|
|
ESSIV_sha256_init256 (ukey, essivhash);
|
|
|
|
u32 ks1[60];
|
|
u32 ks2[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, essivhash, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
luks_decrypt_sector_aes_cbc_essiv256 (luks_bufs->ct_buf, pt_buf, ks1, ks2, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
else if (cipher_mode == HC_LUKS_CIPHER_MODE_CBC_PLAIN)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_128)
|
|
{
|
|
u32 ukey[4];
|
|
|
|
ukey[0] = mk[0];
|
|
ukey[1] = mk[1];
|
|
ukey[2] = mk[2];
|
|
ukey[3] = mk[3];
|
|
|
|
u32 ks1[44];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
luks_decrypt_sector_aes_cbc_plain128 (luks_bufs->ct_buf, pt_buf, ks1, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey[8];
|
|
|
|
ukey[0] = mk[0];
|
|
ukey[1] = mk[1];
|
|
ukey[2] = mk[2];
|
|
ukey[3] = mk[3];
|
|
ukey[4] = mk[4];
|
|
ukey[5] = mk[5];
|
|
ukey[6] = mk[6];
|
|
ukey[7] = mk[7];
|
|
|
|
u32 ks1[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
luks_decrypt_sector_aes_cbc_plain256 (luks_bufs->ct_buf, pt_buf, ks1, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
else if (cipher_mode == HC_LUKS_CIPHER_MODE_XTS_PLAIN)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey1[4];
|
|
|
|
ukey1[0] = mk[0];
|
|
ukey1[1] = mk[1];
|
|
ukey1[2] = mk[2];
|
|
ukey1[3] = mk[3];
|
|
|
|
u32 ukey2[4];
|
|
|
|
ukey2[0] = mk[4];
|
|
ukey2[1] = mk[5];
|
|
ukey2[2] = mk[6];
|
|
ukey2[3] = mk[7];
|
|
|
|
u32 ks1[44];
|
|
u32 ks2[44];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey1, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes128_set_encrypt_key (ks2, ukey2, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
luks_decrypt_sector_aes_xts_plain256 (luks_bufs->ct_buf, pt_buf, ks1, ks2, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_512)
|
|
{
|
|
u32 ukey1[8];
|
|
|
|
ukey1[0] = mk[ 0];
|
|
ukey1[1] = mk[ 1];
|
|
ukey1[2] = mk[ 2];
|
|
ukey1[3] = mk[ 3];
|
|
ukey1[4] = mk[ 4];
|
|
ukey1[5] = mk[ 5];
|
|
ukey1[6] = mk[ 6];
|
|
ukey1[7] = mk[ 7];
|
|
|
|
u32 ukey2[8];
|
|
|
|
ukey2[0] = mk[ 8];
|
|
ukey2[1] = mk[ 9];
|
|
ukey2[2] = mk[10];
|
|
ukey2[3] = mk[11];
|
|
ukey2[4] = mk[12];
|
|
ukey2[5] = mk[13];
|
|
ukey2[6] = mk[14];
|
|
ukey2[7] = mk[15];
|
|
|
|
u32 ks1[60];
|
|
u32 ks2[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey1, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, ukey2, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
luks_decrypt_sector_aes_xts_plain512 (luks_bufs->ct_buf, pt_buf, ks1, ks2, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void luks_af_sha512_then_aes_decrypt (__global luks_t *luks_bufs, __global luks_tmp_t *tmps, u32 *pt_buf, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
const u32 key_size = luks_bufs->key_size;
|
|
const u32 cipher_mode = luks_bufs->cipher_mode;
|
|
|
|
#define BITS_PER_AF (key_size * LUKS_STRIPES)
|
|
#define BITS_PER_SECTOR (512 * 8)
|
|
#define SECTOR_PER_AF (BITS_PER_AF / BITS_PER_SECTOR)
|
|
#define BLOCKS_PER_SECTOR (512 / 16)
|
|
#define OFFSET_PER_BLOCK (16 / 4)
|
|
#define OFFSET_PER_SECTOR (BLOCKS_PER_SECTOR * OFFSET_PER_BLOCK)
|
|
|
|
// decrypt AF data and do the AF merge inline
|
|
|
|
u32 mk[16] = { 0 };
|
|
|
|
if (cipher_mode == HC_LUKS_CIPHER_MODE_CBC_ESSIV)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_128)
|
|
{
|
|
u32 ukey[4];
|
|
|
|
ukey[0] = swap32_S (tmps->out32[1]);
|
|
ukey[1] = swap32_S (tmps->out32[0]);
|
|
ukey[2] = swap32_S (tmps->out32[3]);
|
|
ukey[3] = swap32_S (tmps->out32[2]);
|
|
|
|
u32 essivhash[8];
|
|
|
|
ESSIV_sha256_init128 (ukey, essivhash);
|
|
|
|
u32 ks1[44];
|
|
u32 ks2[60];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, essivhash, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_cbc_essiv128_mk_sha512 (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_cbc_essiv128_mk_sha512_final (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey[8];
|
|
|
|
ukey[0] = swap32_S (tmps->out32[1]);
|
|
ukey[1] = swap32_S (tmps->out32[0]);
|
|
ukey[2] = swap32_S (tmps->out32[3]);
|
|
ukey[3] = swap32_S (tmps->out32[2]);
|
|
ukey[4] = swap32_S (tmps->out32[5]);
|
|
ukey[5] = swap32_S (tmps->out32[4]);
|
|
ukey[6] = swap32_S (tmps->out32[7]);
|
|
ukey[7] = swap32_S (tmps->out32[6]);
|
|
|
|
u32 essivhash[8];
|
|
|
|
ESSIV_sha256_init256 (ukey, essivhash);
|
|
|
|
u32 ks1[60];
|
|
u32 ks2[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, essivhash, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_cbc_essiv256_mk_sha512 (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_cbc_essiv256_mk_sha512_final (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
else if (cipher_mode == HC_LUKS_CIPHER_MODE_CBC_PLAIN)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_128)
|
|
{
|
|
u32 ukey[4];
|
|
|
|
ukey[0] = swap32_S (tmps->out32[1]);
|
|
ukey[1] = swap32_S (tmps->out32[0]);
|
|
ukey[2] = swap32_S (tmps->out32[3]);
|
|
ukey[3] = swap32_S (tmps->out32[2]);
|
|
|
|
u32 ks1[44];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_cbc_plain128_mk_sha512 (luks_bufs->af_src_buf + offset, mk, ks1, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_cbc_plain128_mk_sha512_final (luks_bufs->af_src_buf + offset, mk, ks1, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey[8];
|
|
|
|
ukey[0] = swap32_S (tmps->out32[1]);
|
|
ukey[1] = swap32_S (tmps->out32[0]);
|
|
ukey[2] = swap32_S (tmps->out32[3]);
|
|
ukey[3] = swap32_S (tmps->out32[2]);
|
|
ukey[4] = swap32_S (tmps->out32[5]);
|
|
ukey[5] = swap32_S (tmps->out32[4]);
|
|
ukey[6] = swap32_S (tmps->out32[7]);
|
|
ukey[7] = swap32_S (tmps->out32[6]);
|
|
|
|
u32 ks1[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_cbc_plain256_mk_sha512 (luks_bufs->af_src_buf + offset, mk, ks1, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_cbc_plain256_mk_sha512_final (luks_bufs->af_src_buf + offset, mk, ks1, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
else if (cipher_mode == HC_LUKS_CIPHER_MODE_XTS_PLAIN)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey1[4];
|
|
|
|
ukey1[0] = swap32_S (tmps->out32[1]);
|
|
ukey1[1] = swap32_S (tmps->out32[0]);
|
|
ukey1[2] = swap32_S (tmps->out32[3]);
|
|
ukey1[3] = swap32_S (tmps->out32[2]);
|
|
|
|
u32 ukey2[4];
|
|
|
|
ukey2[0] = swap32_S (tmps->out32[5]);
|
|
ukey2[1] = swap32_S (tmps->out32[4]);
|
|
ukey2[2] = swap32_S (tmps->out32[7]);
|
|
ukey2[3] = swap32_S (tmps->out32[6]);
|
|
|
|
u32 ks1[44];
|
|
u32 ks2[44];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey1, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes128_set_encrypt_key (ks2, ukey2, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_xts_plain256_mk_sha512 (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_xts_plain256_mk_sha512_final (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_512)
|
|
{
|
|
u32 ukey1[8];
|
|
|
|
ukey1[0] = swap32_S (tmps->out32[ 1]);
|
|
ukey1[1] = swap32_S (tmps->out32[ 0]);
|
|
ukey1[2] = swap32_S (tmps->out32[ 3]);
|
|
ukey1[3] = swap32_S (tmps->out32[ 2]);
|
|
ukey1[4] = swap32_S (tmps->out32[ 5]);
|
|
ukey1[5] = swap32_S (tmps->out32[ 4]);
|
|
ukey1[6] = swap32_S (tmps->out32[ 7]);
|
|
ukey1[7] = swap32_S (tmps->out32[ 6]);
|
|
|
|
u32 ukey2[8];
|
|
|
|
ukey2[0] = swap32_S (tmps->out32[ 9]);
|
|
ukey2[1] = swap32_S (tmps->out32[ 8]);
|
|
ukey2[2] = swap32_S (tmps->out32[11]);
|
|
ukey2[3] = swap32_S (tmps->out32[10]);
|
|
ukey2[4] = swap32_S (tmps->out32[13]);
|
|
ukey2[5] = swap32_S (tmps->out32[12]);
|
|
ukey2[6] = swap32_S (tmps->out32[15]);
|
|
ukey2[7] = swap32_S (tmps->out32[14]);
|
|
|
|
u32 ks1[60];
|
|
u32 ks2[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey1, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, ukey2, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_xts_plain512_mk_sha512 (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_xts_plain512_mk_sha512_final (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
|
|
// decrypt payload data
|
|
|
|
if (cipher_mode == HC_LUKS_CIPHER_MODE_CBC_ESSIV)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_128)
|
|
{
|
|
u32 ukey[4];
|
|
|
|
ukey[0] = mk[0];
|
|
ukey[1] = mk[1];
|
|
ukey[2] = mk[2];
|
|
ukey[3] = mk[3];
|
|
|
|
u32 essivhash[8];
|
|
|
|
ESSIV_sha256_init128 (ukey, essivhash);
|
|
|
|
u32 ks1[44];
|
|
u32 ks2[60];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, essivhash, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
luks_decrypt_sector_aes_cbc_essiv128 (luks_bufs->ct_buf, pt_buf, ks1, ks2, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey[8];
|
|
|
|
ukey[0] = mk[0];
|
|
ukey[1] = mk[1];
|
|
ukey[2] = mk[2];
|
|
ukey[3] = mk[3];
|
|
ukey[4] = mk[4];
|
|
ukey[5] = mk[5];
|
|
ukey[6] = mk[6];
|
|
ukey[7] = mk[7];
|
|
|
|
u32 essivhash[8];
|
|
|
|
ESSIV_sha256_init256 (ukey, essivhash);
|
|
|
|
u32 ks1[60];
|
|
u32 ks2[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, essivhash, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
luks_decrypt_sector_aes_cbc_essiv256 (luks_bufs->ct_buf, pt_buf, ks1, ks2, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
else if (cipher_mode == HC_LUKS_CIPHER_MODE_CBC_PLAIN)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_128)
|
|
{
|
|
u32 ukey[4];
|
|
|
|
ukey[0] = mk[0];
|
|
ukey[1] = mk[1];
|
|
ukey[2] = mk[2];
|
|
ukey[3] = mk[3];
|
|
|
|
u32 ks1[44];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
luks_decrypt_sector_aes_cbc_plain128 (luks_bufs->ct_buf, pt_buf, ks1, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey[8];
|
|
|
|
ukey[0] = mk[0];
|
|
ukey[1] = mk[1];
|
|
ukey[2] = mk[2];
|
|
ukey[3] = mk[3];
|
|
ukey[4] = mk[4];
|
|
ukey[5] = mk[5];
|
|
ukey[6] = mk[6];
|
|
ukey[7] = mk[7];
|
|
|
|
u32 ks1[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
luks_decrypt_sector_aes_cbc_plain256 (luks_bufs->ct_buf, pt_buf, ks1, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
else if (cipher_mode == HC_LUKS_CIPHER_MODE_XTS_PLAIN)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey1[4];
|
|
|
|
ukey1[0] = mk[0];
|
|
ukey1[1] = mk[1];
|
|
ukey1[2] = mk[2];
|
|
ukey1[3] = mk[3];
|
|
|
|
u32 ukey2[4];
|
|
|
|
ukey2[0] = mk[4];
|
|
ukey2[1] = mk[5];
|
|
ukey2[2] = mk[6];
|
|
ukey2[3] = mk[7];
|
|
|
|
u32 ks1[44];
|
|
u32 ks2[44];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey1, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes128_set_encrypt_key (ks2, ukey2, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
luks_decrypt_sector_aes_xts_plain256 (luks_bufs->ct_buf, pt_buf, ks1, ks2, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_512)
|
|
{
|
|
u32 ukey1[8];
|
|
|
|
ukey1[0] = mk[ 0];
|
|
ukey1[1] = mk[ 1];
|
|
ukey1[2] = mk[ 2];
|
|
ukey1[3] = mk[ 3];
|
|
ukey1[4] = mk[ 4];
|
|
ukey1[5] = mk[ 5];
|
|
ukey1[6] = mk[ 6];
|
|
ukey1[7] = mk[ 7];
|
|
|
|
u32 ukey2[8];
|
|
|
|
ukey2[0] = mk[ 8];
|
|
ukey2[1] = mk[ 9];
|
|
ukey2[2] = mk[10];
|
|
ukey2[3] = mk[11];
|
|
ukey2[4] = mk[12];
|
|
ukey2[5] = mk[13];
|
|
ukey2[6] = mk[14];
|
|
ukey2[7] = mk[15];
|
|
|
|
u32 ks1[60];
|
|
u32 ks2[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey1, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, ukey2, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
luks_decrypt_sector_aes_xts_plain512 (luks_bufs->ct_buf, pt_buf, ks1, ks2, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void luks_af_ripemd160_then_aes_decrypt (__global luks_t *luks_bufs, __global luks_tmp_t *tmps, u32 *pt_buf, SHM_TYPE u32 *s_te0, SHM_TYPE u32 *s_te1, SHM_TYPE u32 *s_te2, SHM_TYPE u32 *s_te3, SHM_TYPE u32 *s_te4, SHM_TYPE u32 *s_td0, SHM_TYPE u32 *s_td1, SHM_TYPE u32 *s_td2, SHM_TYPE u32 *s_td3, SHM_TYPE u32 *s_td4)
|
|
{
|
|
const u32 key_size = luks_bufs->key_size;
|
|
const u32 cipher_mode = luks_bufs->cipher_mode;
|
|
|
|
#define BITS_PER_AF (key_size * LUKS_STRIPES)
|
|
#define BITS_PER_SECTOR (512 * 8)
|
|
#define SECTOR_PER_AF (BITS_PER_AF / BITS_PER_SECTOR)
|
|
#define BLOCKS_PER_SECTOR (512 / 16)
|
|
#define OFFSET_PER_BLOCK (16 / 4)
|
|
#define OFFSET_PER_SECTOR (BLOCKS_PER_SECTOR * OFFSET_PER_BLOCK)
|
|
|
|
// decrypt AF data and do the AF merge inline
|
|
|
|
u32 mk[16] = { 0 };
|
|
|
|
if (cipher_mode == HC_LUKS_CIPHER_MODE_CBC_ESSIV)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_128)
|
|
{
|
|
u32 ukey[4];
|
|
|
|
ukey[0] = tmps->out32[0];
|
|
ukey[1] = tmps->out32[1];
|
|
ukey[2] = tmps->out32[2];
|
|
ukey[3] = tmps->out32[3];
|
|
|
|
u32 essivhash[8];
|
|
|
|
ESSIV_sha256_init128 (ukey, essivhash);
|
|
|
|
u32 ks1[44];
|
|
u32 ks2[60];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, essivhash, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_cbc_essiv128_mk_ripemd160 (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_cbc_essiv128_mk_ripemd160_final (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey[8];
|
|
|
|
ukey[0] = tmps->out32[0];
|
|
ukey[1] = tmps->out32[1];
|
|
ukey[2] = tmps->out32[2];
|
|
ukey[3] = tmps->out32[3];
|
|
ukey[4] = tmps->out32[4];
|
|
ukey[5] = tmps->out32[5];
|
|
ukey[6] = tmps->out32[6];
|
|
ukey[7] = tmps->out32[7];
|
|
|
|
u32 essivhash[8];
|
|
|
|
ESSIV_sha256_init256 (ukey, essivhash);
|
|
|
|
u32 ks1[60];
|
|
u32 ks2[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, essivhash, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_cbc_essiv256_mk_ripemd160 (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_cbc_essiv256_mk_ripemd160_final (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
else if (cipher_mode == HC_LUKS_CIPHER_MODE_CBC_PLAIN)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_128)
|
|
{
|
|
u32 ukey[4];
|
|
|
|
ukey[0] = tmps->out32[0];
|
|
ukey[1] = tmps->out32[1];
|
|
ukey[2] = tmps->out32[2];
|
|
ukey[3] = tmps->out32[3];
|
|
|
|
u32 ks1[44];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_cbc_plain128_mk_ripemd160 (luks_bufs->af_src_buf + offset, mk, ks1, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_cbc_plain128_mk_ripemd160_final (luks_bufs->af_src_buf + offset, mk, ks1, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey[8];
|
|
|
|
ukey[0] = tmps->out32[0];
|
|
ukey[1] = tmps->out32[1];
|
|
ukey[2] = tmps->out32[2];
|
|
ukey[3] = tmps->out32[3];
|
|
ukey[4] = tmps->out32[4];
|
|
ukey[5] = tmps->out32[5];
|
|
ukey[6] = tmps->out32[6];
|
|
ukey[7] = tmps->out32[7];
|
|
|
|
u32 ks1[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_cbc_plain256_mk_ripemd160 (luks_bufs->af_src_buf + offset, mk, ks1, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_cbc_plain256_mk_ripemd160_final (luks_bufs->af_src_buf + offset, mk, ks1, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
else if (cipher_mode == HC_LUKS_CIPHER_MODE_XTS_PLAIN)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey1[4];
|
|
|
|
ukey1[0] = tmps->out32[0];
|
|
ukey1[1] = tmps->out32[1];
|
|
ukey1[2] = tmps->out32[2];
|
|
ukey1[3] = tmps->out32[3];
|
|
|
|
u32 ukey2[4];
|
|
|
|
ukey2[0] = tmps->out32[4];
|
|
ukey2[1] = tmps->out32[5];
|
|
ukey2[2] = tmps->out32[6];
|
|
ukey2[3] = tmps->out32[7];
|
|
|
|
u32 ks1[44];
|
|
u32 ks2[44];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey1, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes128_set_encrypt_key (ks2, ukey2, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_xts_plain256_mk_ripemd160 (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_xts_plain256_mk_ripemd160_final (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_512)
|
|
{
|
|
u32 ukey1[8];
|
|
|
|
ukey1[0] = tmps->out32[ 0];
|
|
ukey1[1] = tmps->out32[ 1];
|
|
ukey1[2] = tmps->out32[ 2];
|
|
ukey1[3] = tmps->out32[ 3];
|
|
ukey1[4] = tmps->out32[ 4];
|
|
ukey1[5] = tmps->out32[ 5];
|
|
ukey1[6] = tmps->out32[ 6];
|
|
ukey1[7] = tmps->out32[ 7];
|
|
|
|
u32 ukey2[8];
|
|
|
|
ukey2[0] = tmps->out32[ 8];
|
|
ukey2[1] = tmps->out32[ 9];
|
|
ukey2[2] = tmps->out32[10];
|
|
ukey2[3] = tmps->out32[11];
|
|
ukey2[4] = tmps->out32[12];
|
|
ukey2[5] = tmps->out32[13];
|
|
ukey2[6] = tmps->out32[14];
|
|
ukey2[7] = tmps->out32[15];
|
|
|
|
u32 ks1[60];
|
|
u32 ks2[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey1, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, ukey2, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
int sector = 0;
|
|
int offset = 0;
|
|
|
|
for (sector = 0; sector < SECTOR_PER_AF - 1; sector++, offset += OFFSET_PER_SECTOR)
|
|
{
|
|
luks_decrypt_sector_aes_xts_plain512_mk_ripemd160 (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
|
|
luks_decrypt_sector_aes_xts_plain512_mk_ripemd160_final (luks_bufs->af_src_buf + offset, mk, ks1, ks2, sector, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
|
|
// decrypt payload data
|
|
|
|
if (cipher_mode == HC_LUKS_CIPHER_MODE_CBC_ESSIV)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_128)
|
|
{
|
|
u32 ukey[4];
|
|
|
|
ukey[0] = mk[0];
|
|
ukey[1] = mk[1];
|
|
ukey[2] = mk[2];
|
|
ukey[3] = mk[3];
|
|
|
|
u32 essivhash[8];
|
|
|
|
ESSIV_sha256_init128 (ukey, essivhash);
|
|
|
|
u32 ks1[44];
|
|
u32 ks2[60];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, essivhash, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
luks_decrypt_sector_aes_cbc_essiv128 (luks_bufs->ct_buf, pt_buf, ks1, ks2, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey[8];
|
|
|
|
ukey[0] = mk[0];
|
|
ukey[1] = mk[1];
|
|
ukey[2] = mk[2];
|
|
ukey[3] = mk[3];
|
|
ukey[4] = mk[4];
|
|
ukey[5] = mk[5];
|
|
ukey[6] = mk[6];
|
|
ukey[7] = mk[7];
|
|
|
|
u32 essivhash[8];
|
|
|
|
ESSIV_sha256_init256 (ukey, essivhash);
|
|
|
|
u32 ks1[60];
|
|
u32 ks2[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, essivhash, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
luks_decrypt_sector_aes_cbc_essiv256 (luks_bufs->ct_buf, pt_buf, ks1, ks2, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
else if (cipher_mode == HC_LUKS_CIPHER_MODE_CBC_PLAIN)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_128)
|
|
{
|
|
u32 ukey[4];
|
|
|
|
ukey[0] = mk[0];
|
|
ukey[1] = mk[1];
|
|
ukey[2] = mk[2];
|
|
ukey[3] = mk[3];
|
|
|
|
u32 ks1[44];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
luks_decrypt_sector_aes_cbc_plain128 (luks_bufs->ct_buf, pt_buf, ks1, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey[8];
|
|
|
|
ukey[0] = mk[0];
|
|
ukey[1] = mk[1];
|
|
ukey[2] = mk[2];
|
|
ukey[3] = mk[3];
|
|
ukey[4] = mk[4];
|
|
ukey[5] = mk[5];
|
|
ukey[6] = mk[6];
|
|
ukey[7] = mk[7];
|
|
|
|
u32 ks1[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
|
|
luks_decrypt_sector_aes_cbc_plain256 (luks_bufs->ct_buf, pt_buf, ks1, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
else if (cipher_mode == HC_LUKS_CIPHER_MODE_XTS_PLAIN)
|
|
{
|
|
if (key_size == HC_LUKS_KEY_SIZE_256)
|
|
{
|
|
u32 ukey1[4];
|
|
|
|
ukey1[0] = mk[0];
|
|
ukey1[1] = mk[1];
|
|
ukey1[2] = mk[2];
|
|
ukey1[3] = mk[3];
|
|
|
|
u32 ukey2[4];
|
|
|
|
ukey2[0] = mk[4];
|
|
ukey2[1] = mk[5];
|
|
ukey2[2] = mk[6];
|
|
ukey2[3] = mk[7];
|
|
|
|
u32 ks1[44];
|
|
u32 ks2[44];
|
|
|
|
aes128_set_decrypt_key (ks1, ukey1, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes128_set_encrypt_key (ks2, ukey2, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
luks_decrypt_sector_aes_xts_plain256 (luks_bufs->ct_buf, pt_buf, ks1, ks2, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
else if (key_size == HC_LUKS_KEY_SIZE_512)
|
|
{
|
|
u32 ukey1[8];
|
|
|
|
ukey1[0] = mk[ 0];
|
|
ukey1[1] = mk[ 1];
|
|
ukey1[2] = mk[ 2];
|
|
ukey1[3] = mk[ 3];
|
|
ukey1[4] = mk[ 4];
|
|
ukey1[5] = mk[ 5];
|
|
ukey1[6] = mk[ 6];
|
|
ukey1[7] = mk[ 7];
|
|
|
|
u32 ukey2[8];
|
|
|
|
ukey2[0] = mk[ 8];
|
|
ukey2[1] = mk[ 9];
|
|
ukey2[2] = mk[10];
|
|
ukey2[3] = mk[11];
|
|
ukey2[4] = mk[12];
|
|
ukey2[5] = mk[13];
|
|
ukey2[6] = mk[14];
|
|
ukey2[7] = mk[15];
|
|
|
|
u32 ks1[60];
|
|
u32 ks2[60];
|
|
|
|
aes256_set_decrypt_key (ks1, ukey1, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
aes256_set_encrypt_key (ks2, ukey2, s_te0, s_te1, s_te2, s_te3, s_te4);
|
|
|
|
luks_decrypt_sector_aes_xts_plain512 (luks_bufs->ct_buf, pt_buf, ks1, ks2, 0, s_te0, s_te1, s_te2, s_te3, s_te4, s_td0, s_td1, s_td2, s_td3, s_td4);
|
|
}
|
|
}
|
|
}
|