2012-06-18 14:07:19 +03:00
/*
* Shared glue code for 128 bit block ciphers
*
2013-04-08 21:50:55 +03:00
* Copyright © 2012 - 2013 Jussi Kivilinna < jussi . kivilinna @ iki . fi >
2012-06-18 14:07:19 +03:00
*
* CBC & ECB parts based on code ( crypto / cbc . c , ecb . c ) by :
* Copyright ( c ) 2006 Herbert Xu < herbert @ gondor . apana . org . au >
* CTR part based on code ( crypto / ctr . c ) by :
* ( C ) Copyright IBM Corp . 2007 - Joy Latten < latten @ us . ibm . com >
*
* This program is free software ; you can redistribute it and / or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation ; either version 2 of the License , or
* ( at your option ) any later version .
*
* This program is distributed in the hope that it will be useful ,
* but WITHOUT ANY WARRANTY ; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE . See the
* GNU General Public License for more details .
*
* You should have received a copy of the GNU General Public License
* along with this program ; if not , write to the Free Software
* Foundation , Inc . , 59 Temple Place , Suite 330 , Boston , MA 02111 - 1307
* USA
*
*/
# include <linux/module.h>
# include <crypto/b128ops.h>
2017-04-02 21:19:15 +02:00
# include <crypto/gf128mul.h>
2016-11-22 20:08:29 +08:00
# include <crypto/internal/skcipher.h>
2012-06-18 14:07:19 +03:00
# include <crypto/xts.h>
# include <asm/crypto/glue_helper.h>
2018-02-19 23:48:00 -08:00
int glue_ecb_req_128bit ( const struct common_glue_ctx * gctx ,
struct skcipher_request * req )
{
void * ctx = crypto_skcipher_ctx ( crypto_skcipher_reqtfm ( req ) ) ;
const unsigned int bsize = 128 / 8 ;
struct skcipher_walk walk ;
bool fpu_enabled = false ;
unsigned int nbytes ;
int err ;
err = skcipher_walk_virt ( & walk , req , false ) ;
while ( ( nbytes = walk . nbytes ) ) {
const u8 * src = walk . src . virt . addr ;
u8 * dst = walk . dst . virt . addr ;
unsigned int func_bytes ;
unsigned int i ;
2018-02-19 23:48:27 -08:00
fpu_enabled = glue_fpu_begin ( bsize , gctx - > fpu_blocks_limit ,
& walk , fpu_enabled , nbytes ) ;
2018-02-19 23:48:00 -08:00
for ( i = 0 ; i < gctx - > num_funcs ; i + + ) {
func_bytes = bsize * gctx - > funcs [ i ] . num_blocks ;
if ( nbytes < func_bytes )
continue ;
/* Process multi-block batch */
do {
gctx - > funcs [ i ] . fn_u . ecb ( ctx , dst , src ) ;
src + = func_bytes ;
dst + = func_bytes ;
nbytes - = func_bytes ;
} while ( nbytes > = func_bytes ) ;
if ( nbytes < bsize )
break ;
}
err = skcipher_walk_done ( & walk , nbytes ) ;
}
glue_fpu_end ( fpu_enabled ) ;
return err ;
}
EXPORT_SYMBOL_GPL ( glue_ecb_req_128bit ) ;
int glue_cbc_encrypt_req_128bit ( const common_glue_func_t fn ,
struct skcipher_request * req )
{
void * ctx = crypto_skcipher_ctx ( crypto_skcipher_reqtfm ( req ) ) ;
const unsigned int bsize = 128 / 8 ;
struct skcipher_walk walk ;
unsigned int nbytes ;
int err ;
err = skcipher_walk_virt ( & walk , req , false ) ;
while ( ( nbytes = walk . nbytes ) ) {
const u128 * src = ( u128 * ) walk . src . virt . addr ;
u128 * dst = ( u128 * ) walk . dst . virt . addr ;
u128 * iv = ( u128 * ) walk . iv ;
do {
u128_xor ( dst , src , iv ) ;
fn ( ctx , ( u8 * ) dst , ( u8 * ) dst ) ;
iv = dst ;
src + + ;
dst + + ;
nbytes - = bsize ;
} while ( nbytes > = bsize ) ;
* ( u128 * ) walk . iv = * iv ;
err = skcipher_walk_done ( & walk , nbytes ) ;
}
return err ;
}
EXPORT_SYMBOL_GPL ( glue_cbc_encrypt_req_128bit ) ;
int glue_cbc_decrypt_req_128bit ( const struct common_glue_ctx * gctx ,
struct skcipher_request * req )
{
void * ctx = crypto_skcipher_ctx ( crypto_skcipher_reqtfm ( req ) ) ;
const unsigned int bsize = 128 / 8 ;
struct skcipher_walk walk ;
bool fpu_enabled = false ;
unsigned int nbytes ;
int err ;
err = skcipher_walk_virt ( & walk , req , false ) ;
while ( ( nbytes = walk . nbytes ) ) {
const u128 * src = walk . src . virt . addr ;
u128 * dst = walk . dst . virt . addr ;
unsigned int func_bytes , num_blocks ;
unsigned int i ;
u128 last_iv ;
2018-02-19 23:48:27 -08:00
fpu_enabled = glue_fpu_begin ( bsize , gctx - > fpu_blocks_limit ,
& walk , fpu_enabled , nbytes ) ;
2018-02-19 23:48:00 -08:00
/* Start of the last block. */
src + = nbytes / bsize - 1 ;
dst + = nbytes / bsize - 1 ;
last_iv = * src ;
for ( i = 0 ; i < gctx - > num_funcs ; i + + ) {
num_blocks = gctx - > funcs [ i ] . num_blocks ;
func_bytes = bsize * num_blocks ;
if ( nbytes < func_bytes )
continue ;
/* Process multi-block batch */
do {
src - = num_blocks - 1 ;
dst - = num_blocks - 1 ;
gctx - > funcs [ i ] . fn_u . cbc ( ctx , dst , src ) ;
nbytes - = func_bytes ;
if ( nbytes < bsize )
goto done ;
u128_xor ( dst , dst , - - src ) ;
dst - - ;
} while ( nbytes > = func_bytes ) ;
}
done :
u128_xor ( dst , dst , ( u128 * ) walk . iv ) ;
* ( u128 * ) walk . iv = last_iv ;
err = skcipher_walk_done ( & walk , nbytes ) ;
}
glue_fpu_end ( fpu_enabled ) ;
return err ;
}
EXPORT_SYMBOL_GPL ( glue_cbc_decrypt_req_128bit ) ;
int glue_ctr_req_128bit ( const struct common_glue_ctx * gctx ,
struct skcipher_request * req )
{
void * ctx = crypto_skcipher_ctx ( crypto_skcipher_reqtfm ( req ) ) ;
const unsigned int bsize = 128 / 8 ;
struct skcipher_walk walk ;
bool fpu_enabled = false ;
unsigned int nbytes ;
int err ;
err = skcipher_walk_virt ( & walk , req , false ) ;
while ( ( nbytes = walk . nbytes ) > = bsize ) {
const u128 * src = walk . src . virt . addr ;
u128 * dst = walk . dst . virt . addr ;
unsigned int func_bytes , num_blocks ;
unsigned int i ;
le128 ctrblk ;
2018-02-19 23:48:27 -08:00
fpu_enabled = glue_fpu_begin ( bsize , gctx - > fpu_blocks_limit ,
& walk , fpu_enabled , nbytes ) ;
2018-02-19 23:48:00 -08:00
be128_to_le128 ( & ctrblk , ( be128 * ) walk . iv ) ;
for ( i = 0 ; i < gctx - > num_funcs ; i + + ) {
num_blocks = gctx - > funcs [ i ] . num_blocks ;
func_bytes = bsize * num_blocks ;
if ( nbytes < func_bytes )
continue ;
/* Process multi-block batch */
do {
gctx - > funcs [ i ] . fn_u . ctr ( ctx , dst , src , & ctrblk ) ;
src + = num_blocks ;
dst + = num_blocks ;
nbytes - = func_bytes ;
} while ( nbytes > = func_bytes ) ;
if ( nbytes < bsize )
break ;
}
le128_to_be128 ( ( be128 * ) walk . iv , & ctrblk ) ;
err = skcipher_walk_done ( & walk , nbytes ) ;
}
glue_fpu_end ( fpu_enabled ) ;
if ( nbytes ) {
le128 ctrblk ;
u128 tmp ;
be128_to_le128 ( & ctrblk , ( be128 * ) walk . iv ) ;
memcpy ( & tmp , walk . src . virt . addr , nbytes ) ;
gctx - > funcs [ gctx - > num_funcs - 1 ] . fn_u . ctr ( ctx , & tmp , & tmp ,
& ctrblk ) ;
memcpy ( walk . dst . virt . addr , & tmp , nbytes ) ;
le128_to_be128 ( ( be128 * ) walk . iv , & ctrblk ) ;
err = skcipher_walk_done ( & walk , 0 ) ;
}
return err ;
}
EXPORT_SYMBOL_GPL ( glue_ctr_req_128bit ) ;
2016-11-22 20:08:29 +08:00
static unsigned int __glue_xts_req_128bit ( const struct common_glue_ctx * gctx ,
void * ctx ,
struct skcipher_walk * walk )
{
const unsigned int bsize = 128 / 8 ;
unsigned int nbytes = walk - > nbytes ;
u128 * src = walk - > src . virt . addr ;
u128 * dst = walk - > dst . virt . addr ;
unsigned int num_blocks , func_bytes ;
unsigned int i ;
/* Process multi-block batch */
for ( i = 0 ; i < gctx - > num_funcs ; i + + ) {
num_blocks = gctx - > funcs [ i ] . num_blocks ;
func_bytes = bsize * num_blocks ;
if ( nbytes > = func_bytes ) {
do {
gctx - > funcs [ i ] . fn_u . xts ( ctx , dst , src ,
walk - > iv ) ;
src + = num_blocks ;
dst + = num_blocks ;
nbytes - = func_bytes ;
} while ( nbytes > = func_bytes ) ;
if ( nbytes < bsize )
goto done ;
}
}
done :
return nbytes ;
}
int glue_xts_req_128bit ( const struct common_glue_ctx * gctx ,
struct skcipher_request * req ,
common_glue_func_t tweak_fn , void * tweak_ctx ,
void * crypt_ctx )
{
const unsigned int bsize = 128 / 8 ;
struct skcipher_walk walk ;
bool fpu_enabled = false ;
unsigned int nbytes ;
int err ;
err = skcipher_walk_virt ( & walk , req , false ) ;
nbytes = walk . nbytes ;
if ( ! nbytes )
return err ;
/* set minimum length to bsize, for tweak_fn */
2018-02-19 23:48:27 -08:00
fpu_enabled = glue_fpu_begin ( bsize , gctx - > fpu_blocks_limit ,
& walk , fpu_enabled ,
nbytes < bsize ? bsize : nbytes ) ;
2016-11-22 20:08:29 +08:00
/* calculate first value of T */
tweak_fn ( tweak_ctx , walk . iv , walk . iv ) ;
while ( nbytes ) {
nbytes = __glue_xts_req_128bit ( gctx , crypt_ctx , & walk ) ;
err = skcipher_walk_done ( & walk , nbytes ) ;
nbytes = walk . nbytes ;
}
glue_fpu_end ( fpu_enabled ) ;
return err ;
}
EXPORT_SYMBOL_GPL ( glue_xts_req_128bit ) ;
2013-04-08 21:50:55 +03:00
void glue_xts_crypt_128bit_one ( void * ctx , u128 * dst , const u128 * src , le128 * iv ,
common_glue_func_t fn )
{
le128 ivblk = * iv ;
/* generate next IV */
2017-04-02 21:19:15 +02:00
gf128mul_x_ble ( iv , & ivblk ) ;
2013-04-08 21:50:55 +03:00
/* CC <- T xor C */
u128_xor ( dst , src , ( u128 * ) & ivblk ) ;
/* PP <- D(Key2,CC) */
fn ( ctx , ( u8 * ) dst , ( u8 * ) dst ) ;
/* P <- T xor PP */
u128_xor ( dst , dst , ( u128 * ) & ivblk ) ;
}
EXPORT_SYMBOL_GPL ( glue_xts_crypt_128bit_one ) ;
2012-06-18 14:07:19 +03:00
MODULE_LICENSE ( " GPL " ) ;