Commit 3b5cf20c authored by Herbert Xu's avatar Herbert Xu

sunrpc: Use skcipher and ahash/shash

This patch replaces uses of blkcipher with skcipher and the long
obsolete hash interface with either shash (for non-SG users) and
ahash.
Signed-off-by: default avatarHerbert Xu <herbert@gondor.apana.org.au>
parent 2731a944
...@@ -36,7 +36,7 @@ ...@@ -36,7 +36,7 @@
* *
*/ */
#include <linux/crypto.h> #include <crypto/skcipher.h>
#include <linux/sunrpc/auth_gss.h> #include <linux/sunrpc/auth_gss.h>
#include <linux/sunrpc/gss_err.h> #include <linux/sunrpc/gss_err.h>
#include <linux/sunrpc/gss_asn1.h> #include <linux/sunrpc/gss_asn1.h>
...@@ -71,10 +71,10 @@ struct gss_krb5_enctype { ...@@ -71,10 +71,10 @@ struct gss_krb5_enctype {
const u32 keyed_cksum; /* is it a keyed cksum? */ const u32 keyed_cksum; /* is it a keyed cksum? */
const u32 keybytes; /* raw key len, in bytes */ const u32 keybytes; /* raw key len, in bytes */
const u32 keylength; /* final key len, in bytes */ const u32 keylength; /* final key len, in bytes */
u32 (*encrypt) (struct crypto_blkcipher *tfm, u32 (*encrypt) (struct crypto_skcipher *tfm,
void *iv, void *in, void *out, void *iv, void *in, void *out,
int length); /* encryption function */ int length); /* encryption function */
u32 (*decrypt) (struct crypto_blkcipher *tfm, u32 (*decrypt) (struct crypto_skcipher *tfm,
void *iv, void *in, void *out, void *iv, void *in, void *out,
int length); /* decryption function */ int length); /* decryption function */
u32 (*mk_key) (const struct gss_krb5_enctype *gk5e, u32 (*mk_key) (const struct gss_krb5_enctype *gk5e,
...@@ -98,12 +98,12 @@ struct krb5_ctx { ...@@ -98,12 +98,12 @@ struct krb5_ctx {
u32 enctype; u32 enctype;
u32 flags; u32 flags;
const struct gss_krb5_enctype *gk5e; /* enctype-specific info */ const struct gss_krb5_enctype *gk5e; /* enctype-specific info */
struct crypto_blkcipher *enc; struct crypto_skcipher *enc;
struct crypto_blkcipher *seq; struct crypto_skcipher *seq;
struct crypto_blkcipher *acceptor_enc; struct crypto_skcipher *acceptor_enc;
struct crypto_blkcipher *initiator_enc; struct crypto_skcipher *initiator_enc;
struct crypto_blkcipher *acceptor_enc_aux; struct crypto_skcipher *acceptor_enc_aux;
struct crypto_blkcipher *initiator_enc_aux; struct crypto_skcipher *initiator_enc_aux;
u8 Ksess[GSS_KRB5_MAX_KEYLEN]; /* session key */ u8 Ksess[GSS_KRB5_MAX_KEYLEN]; /* session key */
u8 cksum[GSS_KRB5_MAX_KEYLEN]; u8 cksum[GSS_KRB5_MAX_KEYLEN];
s32 endtime; s32 endtime;
...@@ -262,24 +262,24 @@ gss_unwrap_kerberos(struct gss_ctx *ctx_id, int offset, ...@@ -262,24 +262,24 @@ gss_unwrap_kerberos(struct gss_ctx *ctx_id, int offset,
u32 u32
krb5_encrypt(struct crypto_blkcipher *key, krb5_encrypt(struct crypto_skcipher *key,
void *iv, void *in, void *out, int length); void *iv, void *in, void *out, int length);
u32 u32
krb5_decrypt(struct crypto_blkcipher *key, krb5_decrypt(struct crypto_skcipher *key,
void *iv, void *in, void *out, int length); void *iv, void *in, void *out, int length);
int int
gss_encrypt_xdr_buf(struct crypto_blkcipher *tfm, struct xdr_buf *outbuf, gss_encrypt_xdr_buf(struct crypto_skcipher *tfm, struct xdr_buf *outbuf,
int offset, struct page **pages); int offset, struct page **pages);
int int
gss_decrypt_xdr_buf(struct crypto_blkcipher *tfm, struct xdr_buf *inbuf, gss_decrypt_xdr_buf(struct crypto_skcipher *tfm, struct xdr_buf *inbuf,
int offset); int offset);
s32 s32
krb5_make_seq_num(struct krb5_ctx *kctx, krb5_make_seq_num(struct krb5_ctx *kctx,
struct crypto_blkcipher *key, struct crypto_skcipher *key,
int direction, int direction,
u32 seqnum, unsigned char *cksum, unsigned char *buf); u32 seqnum, unsigned char *cksum, unsigned char *buf);
...@@ -320,12 +320,12 @@ gss_krb5_aes_decrypt(struct krb5_ctx *kctx, u32 offset, ...@@ -320,12 +320,12 @@ gss_krb5_aes_decrypt(struct krb5_ctx *kctx, u32 offset,
int int
krb5_rc4_setup_seq_key(struct krb5_ctx *kctx, krb5_rc4_setup_seq_key(struct krb5_ctx *kctx,
struct crypto_blkcipher *cipher, struct crypto_skcipher *cipher,
unsigned char *cksum); unsigned char *cksum);
int int
krb5_rc4_setup_enc_key(struct krb5_ctx *kctx, krb5_rc4_setup_enc_key(struct krb5_ctx *kctx,
struct crypto_blkcipher *cipher, struct crypto_skcipher *cipher,
s32 seqnum); s32 seqnum);
void void
gss_krb5_make_confounder(char *p, u32 conflen); gss_krb5_make_confounder(char *p, u32 conflen);
This diff is collapsed.
...@@ -54,9 +54,9 @@ ...@@ -54,9 +54,9 @@
* WARRANTIES OF MERCHANTIBILITY AND FITNESS FOR A PARTICULAR PURPOSE. * WARRANTIES OF MERCHANTIBILITY AND FITNESS FOR A PARTICULAR PURPOSE.
*/ */
#include <crypto/skcipher.h>
#include <linux/err.h> #include <linux/err.h>
#include <linux/types.h> #include <linux/types.h>
#include <linux/crypto.h>
#include <linux/sunrpc/gss_krb5.h> #include <linux/sunrpc/gss_krb5.h>
#include <linux/sunrpc/xdr.h> #include <linux/sunrpc/xdr.h>
#include <linux/lcm.h> #include <linux/lcm.h>
...@@ -147,7 +147,7 @@ u32 krb5_derive_key(const struct gss_krb5_enctype *gk5e, ...@@ -147,7 +147,7 @@ u32 krb5_derive_key(const struct gss_krb5_enctype *gk5e,
size_t blocksize, keybytes, keylength, n; size_t blocksize, keybytes, keylength, n;
unsigned char *inblockdata, *outblockdata, *rawkey; unsigned char *inblockdata, *outblockdata, *rawkey;
struct xdr_netobj inblock, outblock; struct xdr_netobj inblock, outblock;
struct crypto_blkcipher *cipher; struct crypto_skcipher *cipher;
u32 ret = EINVAL; u32 ret = EINVAL;
blocksize = gk5e->blocksize; blocksize = gk5e->blocksize;
...@@ -157,11 +157,11 @@ u32 krb5_derive_key(const struct gss_krb5_enctype *gk5e, ...@@ -157,11 +157,11 @@ u32 krb5_derive_key(const struct gss_krb5_enctype *gk5e,
if ((inkey->len != keylength) || (outkey->len != keylength)) if ((inkey->len != keylength) || (outkey->len != keylength))
goto err_return; goto err_return;
cipher = crypto_alloc_blkcipher(gk5e->encrypt_name, 0, cipher = crypto_alloc_skcipher(gk5e->encrypt_name, 0,
CRYPTO_ALG_ASYNC); CRYPTO_ALG_ASYNC);
if (IS_ERR(cipher)) if (IS_ERR(cipher))
goto err_return; goto err_return;
if (crypto_blkcipher_setkey(cipher, inkey->data, inkey->len)) if (crypto_skcipher_setkey(cipher, inkey->data, inkey->len))
goto err_return; goto err_return;
/* allocate and set up buffers */ /* allocate and set up buffers */
...@@ -238,7 +238,7 @@ u32 krb5_derive_key(const struct gss_krb5_enctype *gk5e, ...@@ -238,7 +238,7 @@ u32 krb5_derive_key(const struct gss_krb5_enctype *gk5e,
memset(inblockdata, 0, blocksize); memset(inblockdata, 0, blocksize);
kfree(inblockdata); kfree(inblockdata);
err_free_cipher: err_free_cipher:
crypto_free_blkcipher(cipher); crypto_free_skcipher(cipher);
err_return: err_return:
return ret; return ret;
} }
......
...@@ -34,6 +34,8 @@ ...@@ -34,6 +34,8 @@
* *
*/ */
#include <crypto/hash.h>
#include <crypto/skcipher.h>
#include <linux/err.h> #include <linux/err.h>
#include <linux/module.h> #include <linux/module.h>
#include <linux/init.h> #include <linux/init.h>
...@@ -42,7 +44,6 @@ ...@@ -42,7 +44,6 @@
#include <linux/sunrpc/auth.h> #include <linux/sunrpc/auth.h>
#include <linux/sunrpc/gss_krb5.h> #include <linux/sunrpc/gss_krb5.h>
#include <linux/sunrpc/xdr.h> #include <linux/sunrpc/xdr.h>
#include <linux/crypto.h>
#include <linux/sunrpc/gss_krb5_enctypes.h> #include <linux/sunrpc/gss_krb5_enctypes.h>
#if IS_ENABLED(CONFIG_SUNRPC_DEBUG) #if IS_ENABLED(CONFIG_SUNRPC_DEBUG)
...@@ -217,7 +218,7 @@ simple_get_netobj(const void *p, const void *end, struct xdr_netobj *res) ...@@ -217,7 +218,7 @@ simple_get_netobj(const void *p, const void *end, struct xdr_netobj *res)
static inline const void * static inline const void *
get_key(const void *p, const void *end, get_key(const void *p, const void *end,
struct krb5_ctx *ctx, struct crypto_blkcipher **res) struct krb5_ctx *ctx, struct crypto_skcipher **res)
{ {
struct xdr_netobj key; struct xdr_netobj key;
int alg; int alg;
...@@ -245,7 +246,7 @@ get_key(const void *p, const void *end, ...@@ -245,7 +246,7 @@ get_key(const void *p, const void *end,
if (IS_ERR(p)) if (IS_ERR(p))
goto out_err; goto out_err;
*res = crypto_alloc_blkcipher(ctx->gk5e->encrypt_name, 0, *res = crypto_alloc_skcipher(ctx->gk5e->encrypt_name, 0,
CRYPTO_ALG_ASYNC); CRYPTO_ALG_ASYNC);
if (IS_ERR(*res)) { if (IS_ERR(*res)) {
printk(KERN_WARNING "gss_kerberos_mech: unable to initialize " printk(KERN_WARNING "gss_kerberos_mech: unable to initialize "
...@@ -253,7 +254,7 @@ get_key(const void *p, const void *end, ...@@ -253,7 +254,7 @@ get_key(const void *p, const void *end,
*res = NULL; *res = NULL;
goto out_err_free_key; goto out_err_free_key;
} }
if (crypto_blkcipher_setkey(*res, key.data, key.len)) { if (crypto_skcipher_setkey(*res, key.data, key.len)) {
printk(KERN_WARNING "gss_kerberos_mech: error setting key for " printk(KERN_WARNING "gss_kerberos_mech: error setting key for "
"crypto algorithm %s\n", ctx->gk5e->encrypt_name); "crypto algorithm %s\n", ctx->gk5e->encrypt_name);
goto out_err_free_tfm; goto out_err_free_tfm;
...@@ -263,7 +264,7 @@ get_key(const void *p, const void *end, ...@@ -263,7 +264,7 @@ get_key(const void *p, const void *end,
return p; return p;
out_err_free_tfm: out_err_free_tfm:
crypto_free_blkcipher(*res); crypto_free_skcipher(*res);
out_err_free_key: out_err_free_key:
kfree(key.data); kfree(key.data);
p = ERR_PTR(-EINVAL); p = ERR_PTR(-EINVAL);
...@@ -335,30 +336,30 @@ gss_import_v1_context(const void *p, const void *end, struct krb5_ctx *ctx) ...@@ -335,30 +336,30 @@ gss_import_v1_context(const void *p, const void *end, struct krb5_ctx *ctx)
return 0; return 0;
out_err_free_key2: out_err_free_key2:
crypto_free_blkcipher(ctx->seq); crypto_free_skcipher(ctx->seq);
out_err_free_key1: out_err_free_key1:
crypto_free_blkcipher(ctx->enc); crypto_free_skcipher(ctx->enc);
out_err_free_mech: out_err_free_mech:
kfree(ctx->mech_used.data); kfree(ctx->mech_used.data);
out_err: out_err:
return PTR_ERR(p); return PTR_ERR(p);
} }
static struct crypto_blkcipher * static struct crypto_skcipher *
context_v2_alloc_cipher(struct krb5_ctx *ctx, const char *cname, u8 *key) context_v2_alloc_cipher(struct krb5_ctx *ctx, const char *cname, u8 *key)
{ {
struct crypto_blkcipher *cp; struct crypto_skcipher *cp;
cp = crypto_alloc_blkcipher(cname, 0, CRYPTO_ALG_ASYNC); cp = crypto_alloc_skcipher(cname, 0, CRYPTO_ALG_ASYNC);
if (IS_ERR(cp)) { if (IS_ERR(cp)) {
dprintk("gss_kerberos_mech: unable to initialize " dprintk("gss_kerberos_mech: unable to initialize "
"crypto algorithm %s\n", cname); "crypto algorithm %s\n", cname);
return NULL; return NULL;
} }
if (crypto_blkcipher_setkey(cp, key, ctx->gk5e->keylength)) { if (crypto_skcipher_setkey(cp, key, ctx->gk5e->keylength)) {
dprintk("gss_kerberos_mech: error setting key for " dprintk("gss_kerberos_mech: error setting key for "
"crypto algorithm %s\n", cname); "crypto algorithm %s\n", cname);
crypto_free_blkcipher(cp); crypto_free_skcipher(cp);
return NULL; return NULL;
} }
return cp; return cp;
...@@ -412,9 +413,9 @@ context_derive_keys_des3(struct krb5_ctx *ctx, gfp_t gfp_mask) ...@@ -412,9 +413,9 @@ context_derive_keys_des3(struct krb5_ctx *ctx, gfp_t gfp_mask)
return 0; return 0;
out_free_enc: out_free_enc:
crypto_free_blkcipher(ctx->enc); crypto_free_skcipher(ctx->enc);
out_free_seq: out_free_seq:
crypto_free_blkcipher(ctx->seq); crypto_free_skcipher(ctx->seq);
out_err: out_err:
return -EINVAL; return -EINVAL;
} }
...@@ -427,18 +428,17 @@ context_derive_keys_des3(struct krb5_ctx *ctx, gfp_t gfp_mask) ...@@ -427,18 +428,17 @@ context_derive_keys_des3(struct krb5_ctx *ctx, gfp_t gfp_mask)
static int static int
context_derive_keys_rc4(struct krb5_ctx *ctx) context_derive_keys_rc4(struct krb5_ctx *ctx)
{ {
struct crypto_hash *hmac; struct crypto_shash *hmac;
char sigkeyconstant[] = "signaturekey"; char sigkeyconstant[] = "signaturekey";
int slen = strlen(sigkeyconstant) + 1; /* include null terminator */ int slen = strlen(sigkeyconstant) + 1; /* include null terminator */
struct hash_desc desc; struct shash_desc *desc;
struct scatterlist sg[1];
int err; int err;
dprintk("RPC: %s: entered\n", __func__); dprintk("RPC: %s: entered\n", __func__);
/* /*
* derive cksum (aka Ksign) key * derive cksum (aka Ksign) key
*/ */
hmac = crypto_alloc_hash(ctx->gk5e->cksum_name, 0, CRYPTO_ALG_ASYNC); hmac = crypto_alloc_shash(ctx->gk5e->cksum_name, 0, 0);
if (IS_ERR(hmac)) { if (IS_ERR(hmac)) {
dprintk("%s: error %ld allocating hash '%s'\n", dprintk("%s: error %ld allocating hash '%s'\n",
__func__, PTR_ERR(hmac), ctx->gk5e->cksum_name); __func__, PTR_ERR(hmac), ctx->gk5e->cksum_name);
...@@ -446,37 +446,40 @@ context_derive_keys_rc4(struct krb5_ctx *ctx) ...@@ -446,37 +446,40 @@ context_derive_keys_rc4(struct krb5_ctx *ctx)
goto out_err; goto out_err;
} }
err = crypto_hash_setkey(hmac, ctx->Ksess, ctx->gk5e->keylength); err = crypto_shash_setkey(hmac, ctx->Ksess, ctx->gk5e->keylength);
if (err) if (err)
goto out_err_free_hmac; goto out_err_free_hmac;
sg_init_table(sg, 1);
sg_set_buf(sg, sigkeyconstant, slen);
desc.tfm = hmac; desc = kmalloc(sizeof(*desc), GFP_KERNEL);
desc.flags = 0; if (!desc) {
dprintk("%s: failed to allocate hash descriptor for '%s'\n",
err = crypto_hash_init(&desc); __func__, ctx->gk5e->cksum_name);
if (err) err = -ENOMEM;
goto out_err_free_hmac; goto out_err_free_hmac;
}
desc->tfm = hmac;
desc->flags = 0;
err = crypto_hash_digest(&desc, sg, slen, ctx->cksum); err = crypto_shash_digest(desc, sigkeyconstant, slen, ctx->cksum);
kzfree(desc);
if (err) if (err)
goto out_err_free_hmac; goto out_err_free_hmac;
/* /*
* allocate hash, and blkciphers for data and seqnum encryption * allocate hash, and skciphers for data and seqnum encryption
*/ */
ctx->enc = crypto_alloc_blkcipher(ctx->gk5e->encrypt_name, 0, ctx->enc = crypto_alloc_skcipher(ctx->gk5e->encrypt_name, 0,
CRYPTO_ALG_ASYNC); CRYPTO_ALG_ASYNC);
if (IS_ERR(ctx->enc)) { if (IS_ERR(ctx->enc)) {
err = PTR_ERR(ctx->enc); err = PTR_ERR(ctx->enc);
goto out_err_free_hmac; goto out_err_free_hmac;
} }
ctx->seq = crypto_alloc_blkcipher(ctx->gk5e->encrypt_name, 0, ctx->seq = crypto_alloc_skcipher(ctx->gk5e->encrypt_name, 0,
CRYPTO_ALG_ASYNC); CRYPTO_ALG_ASYNC);
if (IS_ERR(ctx->seq)) { if (IS_ERR(ctx->seq)) {
crypto_free_blkcipher(ctx->enc); crypto_free_skcipher(ctx->enc);
err = PTR_ERR(ctx->seq); err = PTR_ERR(ctx->seq);
goto out_err_free_hmac; goto out_err_free_hmac;
} }
...@@ -486,7 +489,7 @@ context_derive_keys_rc4(struct krb5_ctx *ctx) ...@@ -486,7 +489,7 @@ context_derive_keys_rc4(struct krb5_ctx *ctx)
err = 0; err = 0;
out_err_free_hmac: out_err_free_hmac:
crypto_free_hash(hmac); crypto_free_shash(hmac);
out_err: out_err:
dprintk("RPC: %s: returning %d\n", __func__, err); dprintk("RPC: %s: returning %d\n", __func__, err);
return err; return err;
...@@ -588,7 +591,7 @@ context_derive_keys_new(struct krb5_ctx *ctx, gfp_t gfp_mask) ...@@ -588,7 +591,7 @@ context_derive_keys_new(struct krb5_ctx *ctx, gfp_t gfp_mask)
context_v2_alloc_cipher(ctx, "cbc(aes)", context_v2_alloc_cipher(ctx, "cbc(aes)",
ctx->acceptor_seal); ctx->acceptor_seal);
if (ctx->acceptor_enc_aux == NULL) { if (ctx->acceptor_enc_aux == NULL) {
crypto_free_blkcipher(ctx->initiator_enc_aux); crypto_free_skcipher(ctx->initiator_enc_aux);
goto out_free_acceptor_enc; goto out_free_acceptor_enc;
} }
} }
...@@ -596,9 +599,9 @@ context_derive_keys_new(struct krb5_ctx *ctx, gfp_t gfp_mask) ...@@ -596,9 +599,9 @@ context_derive_keys_new(struct krb5_ctx *ctx, gfp_t gfp_mask)
return 0; return 0;
out_free_acceptor_enc: out_free_acceptor_enc:
crypto_free_blkcipher(ctx->acceptor_enc); crypto_free_skcipher(ctx->acceptor_enc);
out_free_initiator_enc: out_free_initiator_enc:
crypto_free_blkcipher(ctx->initiator_enc); crypto_free_skcipher(ctx->initiator_enc);
out_err: out_err:
return -EINVAL; return -EINVAL;
} }
...@@ -710,12 +713,12 @@ static void ...@@ -710,12 +713,12 @@ static void
gss_delete_sec_context_kerberos(void *internal_ctx) { gss_delete_sec_context_kerberos(void *internal_ctx) {
struct krb5_ctx *kctx = internal_ctx; struct krb5_ctx *kctx = internal_ctx;
crypto_free_blkcipher(kctx->seq); crypto_free_skcipher(kctx->seq);
crypto_free_blkcipher(kctx->enc); crypto_free_skcipher(kctx->enc);
crypto_free_blkcipher(kctx->acceptor_enc); crypto_free_skcipher(kctx->acceptor_enc);
crypto_free_blkcipher(kctx->initiator_enc); crypto_free_skcipher(kctx->initiator_enc);
crypto_free_blkcipher(kctx->acceptor_enc_aux); crypto_free_skcipher(kctx->acceptor_enc_aux);
crypto_free_blkcipher(kctx->initiator_enc_aux); crypto_free_skcipher(kctx->initiator_enc_aux);
kfree(kctx->mech_used.data); kfree(kctx->mech_used.data);
kfree(kctx); kfree(kctx);
} }
......
...@@ -31,9 +31,9 @@ ...@@ -31,9 +31,9 @@
* PERFORMANCE OF THIS SOFTWARE. * PERFORMANCE OF THIS SOFTWARE.
*/ */
#include <crypto/skcipher.h>
#include <linux/types.h> #include <linux/types.h>
#include <linux/sunrpc/gss_krb5.h> #include <linux/sunrpc/gss_krb5.h>
#include <linux/crypto.h>
#if IS_ENABLED(CONFIG_SUNRPC_DEBUG) #if IS_ENABLED(CONFIG_SUNRPC_DEBUG)
# define RPCDBG_FACILITY RPCDBG_AUTH # define RPCDBG_FACILITY RPCDBG_AUTH
...@@ -43,13 +43,13 @@ static s32 ...@@ -43,13 +43,13 @@ static s32
krb5_make_rc4_seq_num(struct krb5_ctx *kctx, int direction, s32 seqnum, krb5_make_rc4_seq_num(struct krb5_ctx *kctx, int direction, s32 seqnum,
unsigned char *cksum, unsigned char *buf) unsigned char *cksum, unsigned char *buf)
{ {
struct crypto_blkcipher *cipher; struct crypto_skcipher *cipher;
unsigned char plain[8]; unsigned char plain[8];
s32 code; s32 code;
dprintk("RPC: %s:\n", __func__); dprintk("RPC: %s:\n", __func__);
cipher = crypto_alloc_blkcipher(kctx->gk5e->encrypt_name, 0, cipher = crypto_alloc_skcipher(kctx->gk5e->encrypt_name, 0,
CRYPTO_ALG_ASYNC); CRYPTO_ALG_ASYNC);
if (IS_ERR(cipher)) if (IS_ERR(cipher))
return PTR_ERR(cipher); return PTR_ERR(cipher);
...@@ -68,12 +68,12 @@ krb5_make_rc4_seq_num(struct krb5_ctx *kctx, int direction, s32 seqnum, ...@@ -68,12 +68,12 @@ krb5_make_rc4_seq_num(struct krb5_ctx *kctx, int direction, s32 seqnum,
code = krb5_encrypt(cipher, cksum, plain, buf, 8); code = krb5_encrypt(cipher, cksum, plain, buf, 8);
out: out:
crypto_free_blkcipher(cipher); crypto_free_skcipher(cipher);
return code; return code;
} }
s32 s32
krb5_make_seq_num(struct krb5_ctx *kctx, krb5_make_seq_num(struct krb5_ctx *kctx,
struct crypto_blkcipher *key, struct crypto_skcipher *key,
int direction, int direction,
u32 seqnum, u32 seqnum,
unsigned char *cksum, unsigned char *buf) unsigned char *cksum, unsigned char *buf)
...@@ -101,13 +101,13 @@ static s32 ...@@ -101,13 +101,13 @@ static s32
krb5_get_rc4_seq_num(struct krb5_ctx *kctx, unsigned char *cksum, krb5_get_rc4_seq_num(struct krb5_ctx *kctx, unsigned char *cksum,
unsigned char *buf, int *direction, s32 *seqnum) unsigned char *buf, int *direction, s32 *seqnum)
{ {
struct crypto_blkcipher *cipher; struct crypto_skcipher *cipher;
unsigned char plain[8]; unsigned char plain[8];
s32 code; s32 code;
dprintk("RPC: %s:\n", __func__); dprintk("RPC: %s:\n", __func__);
cipher = crypto_alloc_blkcipher(kctx->gk5e->encrypt_name, 0, cipher = crypto_alloc_skcipher(kctx->gk5e->encrypt_name, 0,
CRYPTO_ALG_ASYNC); CRYPTO_ALG_ASYNC);
if (IS_ERR(cipher)) if (IS_ERR(cipher))
return PTR_ERR(cipher); return PTR_ERR(cipher);
...@@ -130,7 +130,7 @@ krb5_get_rc4_seq_num(struct krb5_ctx *kctx, unsigned char *cksum, ...@@ -130,7 +130,7 @@ krb5_get_rc4_seq_num(struct krb5_ctx *kctx, unsigned char *cksum,
*seqnum = ((plain[0] << 24) | (plain[1] << 16) | *seqnum = ((plain[0] << 24) | (plain[1] << 16) |
(plain[2] << 8) | (plain[3])); (plain[2] << 8) | (plain[3]));
out: out:
crypto_free_blkcipher(cipher); crypto_free_skcipher(cipher);
return code; return code;
} }
...@@ -142,7 +142,7 @@ krb5_get_seq_num(struct krb5_ctx *kctx, ...@@ -142,7 +142,7 @@ krb5_get_seq_num(struct krb5_ctx *kctx,
{ {
s32 code; s32 code;
unsigned char plain[8]; unsigned char plain[8];
struct crypto_blkcipher *key = kctx->seq; struct crypto_skcipher *key = kctx->seq;
dprintk("RPC: krb5_get_seq_num:\n"); dprintk("RPC: krb5_get_seq_num:\n");
......
...@@ -28,12 +28,12 @@ ...@@ -28,12 +28,12 @@
* SUCH DAMAGES. * SUCH DAMAGES.
*/ */
#include <crypto/skcipher.h>
#include <linux/types.h> #include <linux/types.h>
#include <linux/jiffies.h> #include <linux/jiffies.h>
#include <linux/sunrpc/gss_krb5.h> #include <linux/sunrpc/gss_krb5.h>
#include <linux/random.h> #include <linux/random.h>
#include <linux/pagemap.h> #include <linux/pagemap.h>
#include <linux/crypto.h>
#if IS_ENABLED(CONFIG_SUNRPC_DEBUG) #if IS_ENABLED(CONFIG_SUNRPC_DEBUG)
# define RPCDBG_FACILITY RPCDBG_AUTH # define RPCDBG_FACILITY RPCDBG_AUTH
...@@ -174,7 +174,7 @@ gss_wrap_kerberos_v1(struct krb5_ctx *kctx, int offset, ...@@ -174,7 +174,7 @@ gss_wrap_kerberos_v1(struct krb5_ctx *kctx, int offset,
now = get_seconds(); now = get_seconds();
blocksize = crypto_blkcipher_blocksize(kctx->enc); blocksize = crypto_skcipher_blocksize(kctx->enc);
gss_krb5_add_padding(buf, offset, blocksize); gss_krb5_add_padding(buf, offset, blocksize);
BUG_ON((buf->len - offset) % blocksize); BUG_ON((buf->len - offset) % blocksize);
plainlen = conflen + buf->len - offset; plainlen = conflen + buf->len - offset;
...@@ -239,10 +239,10 @@ gss_wrap_kerberos_v1(struct krb5_ctx *kctx, int offset, ...@@ -239,10 +239,10 @@ gss_wrap_kerberos_v1(struct krb5_ctx *kctx, int offset,
return GSS_S_FAILURE; return GSS_S_FAILURE;
if (kctx->enctype == ENCTYPE_ARCFOUR_HMAC) { if (kctx->enctype == ENCTYPE_ARCFOUR_HMAC) {
struct crypto_blkcipher *cipher; struct crypto_skcipher *cipher;
int err; int err;
cipher = crypto_alloc_blkcipher(kctx->gk5e->encrypt_name, 0, cipher = crypto_alloc_skcipher(kctx->gk5e->encrypt_name, 0,
CRYPTO_ALG_ASYNC); CRYPTO_ALG_ASYNC);
if (IS_ERR(cipher)) if (IS_ERR(cipher))
return GSS_S_FAILURE; return GSS_S_FAILURE;
...@@ -250,7 +250,7 @@ gss_wrap_kerberos_v1(struct krb5_ctx *kctx, int offset, ...@@ -250,7 +250,7 @@ gss_wrap_kerberos_v1(struct krb5_ctx *kctx, int offset,
err = gss_encrypt_xdr_buf(cipher, buf, err = gss_encrypt_xdr_buf(cipher, buf,
offset + headlen - conflen, pages); offset + headlen - conflen, pages);
crypto_free_blkcipher(cipher); crypto_free_skcipher(cipher);
if (err) if (err)
return GSS_S_FAILURE; return GSS_S_FAILURE;
} else { } else {
...@@ -327,18 +327,18 @@ gss_unwrap_kerberos_v1(struct krb5_ctx *kctx, int offset, struct xdr_buf *buf) ...@@ -327,18 +327,18 @@ gss_unwrap_kerberos_v1(struct krb5_ctx *kctx, int offset, struct xdr_buf *buf)
return GSS_S_BAD_SIG; return GSS_S_BAD_SIG;
if (kctx->enctype == ENCTYPE_ARCFOUR_HMAC) { if (kctx->enctype == ENCTYPE_ARCFOUR_HMAC) {
struct crypto_blkcipher *cipher; struct crypto_skcipher *cipher;
int err; int err;
cipher = crypto_alloc_blkcipher(kctx->gk5e->encrypt_name, 0, cipher = crypto_alloc_skcipher(kctx->gk5e->encrypt_name, 0,
CRYPTO_ALG_ASYNC); CRYPTO_ALG_ASYNC);
if (IS_ERR(cipher)) if (IS_ERR(cipher))
return GSS_S_FAILURE; return GSS_S_FAILURE;
krb5_rc4_setup_enc_key(kctx, cipher, seqnum); krb5_rc4_setup_enc_key(kctx, cipher, seqnum);
err = gss_decrypt_xdr_buf(cipher, buf, crypt_offset); err = gss_decrypt_xdr_buf(cipher, buf, crypt_offset);
crypto_free_blkcipher(cipher); crypto_free_skcipher(cipher);
if (err) if (err)
return GSS_S_DEFECTIVE_TOKEN; return GSS_S_DEFECTIVE_TOKEN;
} else { } else {
...@@ -371,7 +371,7 @@ gss_unwrap_kerberos_v1(struct krb5_ctx *kctx, int offset, struct xdr_buf *buf) ...@@ -371,7 +371,7 @@ gss_unwrap_kerberos_v1(struct krb5_ctx *kctx, int offset, struct xdr_buf *buf)
/* Copy the data back to the right position. XXX: Would probably be /* Copy the data back to the right position. XXX: Would probably be
* better to copy and encrypt at the same time. */ * better to copy and encrypt at the same time. */
blocksize = crypto_blkcipher_blocksize(kctx->enc); blocksize = crypto_skcipher_blocksize(kctx->enc);
data_start = ptr + (GSS_KRB5_TOK_HDR_LEN + kctx->gk5e->cksumlength) + data_start = ptr + (GSS_KRB5_TOK_HDR_LEN + kctx->gk5e->cksumlength) +
conflen; conflen;
orig_start = buf->head[0].iov_base + offset; orig_start = buf->head[0].iov_base + offset;
...@@ -473,7 +473,7 @@ gss_wrap_kerberos_v2(struct krb5_ctx *kctx, u32 offset, ...@@ -473,7 +473,7 @@ gss_wrap_kerberos_v2(struct krb5_ctx *kctx, u32 offset,
*ptr++ = 0xff; *ptr++ = 0xff;
be16ptr = (__be16 *)ptr; be16ptr = (__be16 *)ptr;
blocksize = crypto_blkcipher_blocksize(kctx->acceptor_enc); blocksize = crypto_skcipher_blocksize(kctx->acceptor_enc);
*be16ptr++ = 0; *be16ptr++ = 0;
/* "inner" token header always uses 0 for RRC */ /* "inner" token header always uses 0 for RRC */
*be16ptr++ = 0; *be16ptr++ = 0;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment