Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Corentin Labbe | 1324 | 55.61% | 2 | 10.00% |
Steffen Klassert | 723 | 30.37% | 6 | 30.00% |
Tadeusz Struk | 80 | 3.36% | 1 | 5.00% |
Salvatore Benedetto | 80 | 3.36% | 1 | 5.00% |
Giovanni Cabiddu | 78 | 3.28% | 1 | 5.00% |
Mathias Krause | 48 | 2.02% | 2 | 10.00% |
Herbert Xu | 25 | 1.05% | 4 | 20.00% |
David S. Miller | 20 | 0.84% | 1 | 5.00% |
Eric W. Biedermann | 2 | 0.08% | 1 | 5.00% |
Jussi Kivilinna | 1 | 0.04% | 1 | 5.00% |
Total | 2381 | 20 |
// SPDX-License-Identifier: GPL-2.0 /* * Crypto user configuration API. * * Copyright (C) 2017-2018 Corentin Labbe <clabbe@baylibre.com> * */ #include <linux/crypto.h> #include <linux/cryptouser.h> #include <linux/sched.h> #include <net/netlink.h> #include <crypto/internal/skcipher.h> #include <crypto/internal/rng.h> #include <crypto/akcipher.h> #include <crypto/kpp.h> #include <crypto/internal/cryptouser.h> #include "internal.h" #define null_terminated(x) (strnlen(x, sizeof(x)) < sizeof(x)) static DEFINE_MUTEX(crypto_cfg_mutex); extern struct sock *crypto_nlsk; struct crypto_dump_info { struct sk_buff *in_skb; struct sk_buff *out_skb; u32 nlmsg_seq; u16 nlmsg_flags; }; static int crypto_report_aead(struct sk_buff *skb, struct crypto_alg *alg) { struct crypto_stat raead; u64 v64; u32 v32; memset(&raead, 0, sizeof(raead)); strncpy(raead.type, "aead", sizeof(raead.type)); v32 = atomic_read(&alg->encrypt_cnt); raead.stat_encrypt_cnt = v32; v64 = atomic64_read(&alg->encrypt_tlen); raead.stat_encrypt_tlen = v64; v32 = atomic_read(&alg->decrypt_cnt); raead.stat_decrypt_cnt = v32; v64 = atomic64_read(&alg->decrypt_tlen); raead.stat_decrypt_tlen = v64; v32 = atomic_read(&alg->aead_err_cnt); raead.stat_aead_err_cnt = v32; if (nla_put(skb, CRYPTOCFGA_STAT_AEAD, sizeof(struct crypto_stat), &raead)) goto nla_put_failure; return 0; nla_put_failure: return -EMSGSIZE; } static int crypto_report_cipher(struct sk_buff *skb, struct crypto_alg *alg) { struct crypto_stat rcipher; u64 v64; u32 v32; memset(&rcipher, 0, sizeof(rcipher)); strlcpy(rcipher.type, "cipher", sizeof(rcipher.type)); v32 = atomic_read(&alg->encrypt_cnt); rcipher.stat_encrypt_cnt = v32; v64 = atomic64_read(&alg->encrypt_tlen); rcipher.stat_encrypt_tlen = v64; v32 = atomic_read(&alg->decrypt_cnt); rcipher.stat_decrypt_cnt = v32; v64 = atomic64_read(&alg->decrypt_tlen); rcipher.stat_decrypt_tlen = v64; v32 = atomic_read(&alg->cipher_err_cnt); rcipher.stat_cipher_err_cnt = v32; if (nla_put(skb, CRYPTOCFGA_STAT_CIPHER, sizeof(struct crypto_stat), &rcipher)) goto nla_put_failure; return 0; nla_put_failure: return -EMSGSIZE; } static int crypto_report_comp(struct sk_buff *skb, struct crypto_alg *alg) { struct crypto_stat rcomp; u64 v64; u32 v32; memset(&rcomp, 0, sizeof(rcomp)); strlcpy(rcomp.type, "compression", sizeof(rcomp.type)); v32 = atomic_read(&alg->compress_cnt); rcomp.stat_compress_cnt = v32; v64 = atomic64_read(&alg->compress_tlen); rcomp.stat_compress_tlen = v64; v32 = atomic_read(&alg->decompress_cnt); rcomp.stat_decompress_cnt = v32; v64 = atomic64_read(&alg->decompress_tlen); rcomp.stat_decompress_tlen = v64; v32 = atomic_read(&alg->cipher_err_cnt); rcomp.stat_compress_err_cnt = v32; if (nla_put(skb, CRYPTOCFGA_STAT_COMPRESS, sizeof(struct crypto_stat), &rcomp)) goto nla_put_failure; return 0; nla_put_failure: return -EMSGSIZE; } static int crypto_report_acomp(struct sk_buff *skb, struct crypto_alg *alg) { struct crypto_stat racomp; u64 v64; u32 v32; memset(&racomp, 0, sizeof(racomp)); strlcpy(racomp.type, "acomp", sizeof(racomp.type)); v32 = atomic_read(&alg->compress_cnt); racomp.stat_compress_cnt = v32; v64 = atomic64_read(&alg->compress_tlen); racomp.stat_compress_tlen = v64; v32 = atomic_read(&alg->decompress_cnt); racomp.stat_decompress_cnt = v32; v64 = atomic64_read(&alg->decompress_tlen); racomp.stat_decompress_tlen = v64; v32 = atomic_read(&alg->cipher_err_cnt); racomp.stat_compress_err_cnt = v32; if (nla_put(skb, CRYPTOCFGA_STAT_ACOMP, sizeof(struct crypto_stat), &racomp)) goto nla_put_failure; return 0; nla_put_failure: return -EMSGSIZE; } static int crypto_report_akcipher(struct sk_buff *skb, struct crypto_alg *alg) { struct crypto_stat rakcipher; u64 v64; u32 v32; memset(&rakcipher, 0, sizeof(rakcipher)); strncpy(rakcipher.type, "akcipher", sizeof(rakcipher.type)); v32 = atomic_read(&alg->encrypt_cnt); rakcipher.stat_encrypt_cnt = v32; v64 = atomic64_read(&alg->encrypt_tlen); rakcipher.stat_encrypt_tlen = v64; v32 = atomic_read(&alg->decrypt_cnt); rakcipher.stat_decrypt_cnt = v32; v64 = atomic64_read(&alg->decrypt_tlen); rakcipher.stat_decrypt_tlen = v64; v32 = atomic_read(&alg->sign_cnt); rakcipher.stat_sign_cnt = v32; v32 = atomic_read(&alg->verify_cnt); rakcipher.stat_verify_cnt = v32; v32 = atomic_read(&alg->akcipher_err_cnt); rakcipher.stat_akcipher_err_cnt = v32; if (nla_put(skb, CRYPTOCFGA_STAT_AKCIPHER, sizeof(struct crypto_stat), &rakcipher)) goto nla_put_failure; return 0; nla_put_failure: return -EMSGSIZE; } static int crypto_report_kpp(struct sk_buff *skb, struct crypto_alg *alg) { struct crypto_stat rkpp; u32 v; memset(&rkpp, 0, sizeof(rkpp)); strlcpy(rkpp.type, "kpp", sizeof(rkpp.type)); v = atomic_read(&alg->setsecret_cnt); rkpp.stat_setsecret_cnt = v; v = atomic_read(&alg->generate_public_key_cnt); rkpp.stat_generate_public_key_cnt = v; v = atomic_read(&alg->compute_shared_secret_cnt); rkpp.stat_compute_shared_secret_cnt = v; v = atomic_read(&alg->kpp_err_cnt); rkpp.stat_kpp_err_cnt = v; if (nla_put(skb, CRYPTOCFGA_STAT_KPP, sizeof(struct crypto_stat), &rkpp)) goto nla_put_failure; return 0; nla_put_failure: return -EMSGSIZE; } static int crypto_report_ahash(struct sk_buff *skb, struct crypto_alg *alg) { struct crypto_stat rhash; u64 v64; u32 v32; memset(&rhash, 0, sizeof(rhash)); strncpy(rhash.type, "ahash", sizeof(rhash.type)); v32 = atomic_read(&alg->hash_cnt); rhash.stat_hash_cnt = v32; v64 = atomic64_read(&alg->hash_tlen); rhash.stat_hash_tlen = v64; v32 = atomic_read(&alg->hash_err_cnt); rhash.stat_hash_err_cnt = v32; if (nla_put(skb, CRYPTOCFGA_STAT_HASH, sizeof(struct crypto_stat), &rhash)) goto nla_put_failure; return 0; nla_put_failure: return -EMSGSIZE; } static int crypto_report_shash(struct sk_buff *skb, struct crypto_alg *alg) { struct crypto_stat rhash; u64 v64; u32 v32; memset(&rhash, 0, sizeof(rhash)); strncpy(rhash.type, "shash", sizeof(rhash.type)); v32 = atomic_read(&alg->hash_cnt); rhash.stat_hash_cnt = v32; v64 = atomic64_read(&alg->hash_tlen); rhash.stat_hash_tlen = v64; v32 = atomic_read(&alg->hash_err_cnt); rhash.stat_hash_err_cnt = v32; if (nla_put(skb, CRYPTOCFGA_STAT_HASH, sizeof(struct crypto_stat), &rhash)) goto nla_put_failure; return 0; nla_put_failure: return -EMSGSIZE; } static int crypto_report_rng(struct sk_buff *skb, struct crypto_alg *alg) { struct crypto_stat rrng; u64 v64; u32 v32; memset(&rrng, 0, sizeof(rrng)); strncpy(rrng.type, "rng", sizeof(rrng.type)); v32 = atomic_read(&alg->generate_cnt); rrng.stat_generate_cnt = v32; v64 = atomic64_read(&alg->generate_tlen); rrng.stat_generate_tlen = v64; v32 = atomic_read(&alg->seed_cnt); rrng.stat_seed_cnt = v32; v32 = atomic_read(&alg->hash_err_cnt); rrng.stat_rng_err_cnt = v32; if (nla_put(skb, CRYPTOCFGA_STAT_RNG, sizeof(struct crypto_stat), &rrng)) goto nla_put_failure; return 0; nla_put_failure: return -EMSGSIZE; } static int crypto_reportstat_one(struct crypto_alg *alg, struct crypto_user_alg *ualg, struct sk_buff *skb) { memset(ualg, 0, sizeof(*ualg)); strlcpy(ualg->cru_name, alg->cra_name, sizeof(ualg->cru_name)); strlcpy(ualg->cru_driver_name, alg->cra_driver_name, sizeof(ualg->cru_driver_name)); strlcpy(ualg->cru_module_name, module_name(alg->cra_module), sizeof(ualg->cru_module_name)); ualg->cru_type = 0; ualg->cru_mask = 0; ualg->cru_flags = alg->cra_flags; ualg->cru_refcnt = refcount_read(&alg->cra_refcnt); if (nla_put_u32(skb, CRYPTOCFGA_PRIORITY_VAL, alg->cra_priority)) goto nla_put_failure; if (alg->cra_flags & CRYPTO_ALG_LARVAL) { struct crypto_stat rl; memset(&rl, 0, sizeof(rl)); strlcpy(rl.type, "larval", sizeof(rl.type)); if (nla_put(skb, CRYPTOCFGA_STAT_LARVAL, sizeof(struct crypto_stat), &rl)) goto nla_put_failure; goto out; } switch (alg->cra_flags & (CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_LARVAL)) { case CRYPTO_ALG_TYPE_AEAD: if (crypto_report_aead(skb, alg)) goto nla_put_failure; break; case CRYPTO_ALG_TYPE_SKCIPHER: if (crypto_report_cipher(skb, alg)) goto nla_put_failure; break; case CRYPTO_ALG_TYPE_BLKCIPHER: if (crypto_report_cipher(skb, alg)) goto nla_put_failure; break; case CRYPTO_ALG_TYPE_CIPHER: if (crypto_report_cipher(skb, alg)) goto nla_put_failure; break; case CRYPTO_ALG_TYPE_COMPRESS: if (crypto_report_comp(skb, alg)) goto nla_put_failure; break; case CRYPTO_ALG_TYPE_ACOMPRESS: if (crypto_report_acomp(skb, alg)) goto nla_put_failure; break; case CRYPTO_ALG_TYPE_SCOMPRESS: if (crypto_report_acomp(skb, alg)) goto nla_put_failure; break; case CRYPTO_ALG_TYPE_AKCIPHER: if (crypto_report_akcipher(skb, alg)) goto nla_put_failure; break; case CRYPTO_ALG_TYPE_KPP: if (crypto_report_kpp(skb, alg)) goto nla_put_failure; break; case CRYPTO_ALG_TYPE_AHASH: if (crypto_report_ahash(skb, alg)) goto nla_put_failure; break; case CRYPTO_ALG_TYPE_HASH: if (crypto_report_shash(skb, alg)) goto nla_put_failure; break; case CRYPTO_ALG_TYPE_RNG: if (crypto_report_rng(skb, alg)) goto nla_put_failure; break; default: pr_err("ERROR: Unhandled alg %d in %s\n", alg->cra_flags & (CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_LARVAL), __func__); } out: return 0; nla_put_failure: return -EMSGSIZE; } static int crypto_reportstat_alg(struct crypto_alg *alg, struct crypto_dump_info *info) { struct sk_buff *in_skb = info->in_skb; struct sk_buff *skb = info->out_skb; struct nlmsghdr *nlh; struct crypto_user_alg *ualg; int err = 0; nlh = nlmsg_put(skb, NETLINK_CB(in_skb).portid, info->nlmsg_seq, CRYPTO_MSG_GETSTAT, sizeof(*ualg), info->nlmsg_flags); if (!nlh) { err = -EMSGSIZE; goto out; } ualg = nlmsg_data(nlh); err = crypto_reportstat_one(alg, ualg, skb); if (err) { nlmsg_cancel(skb, nlh); goto out; } nlmsg_end(skb, nlh); out: return err; } int crypto_reportstat(struct sk_buff *in_skb, struct nlmsghdr *in_nlh, struct nlattr **attrs) { struct crypto_user_alg *p = nlmsg_data(in_nlh); struct crypto_alg *alg; struct sk_buff *skb; struct crypto_dump_info info; int err; if (!null_terminated(p->cru_name) || !null_terminated(p->cru_driver_name)) return -EINVAL; alg = crypto_alg_match(p, 0); if (!alg) return -ENOENT; err = -ENOMEM; skb = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_ATOMIC); if (!skb) goto drop_alg; info.in_skb = in_skb; info.out_skb = skb; info.nlmsg_seq = in_nlh->nlmsg_seq; info.nlmsg_flags = 0; err = crypto_reportstat_alg(alg, &info); drop_alg: crypto_mod_put(alg); if (err) return err; return nlmsg_unicast(crypto_nlsk, skb, NETLINK_CB(in_skb).portid); } int crypto_dump_reportstat(struct sk_buff *skb, struct netlink_callback *cb) { struct crypto_alg *alg; struct crypto_dump_info info; int err; if (cb->args[0]) goto out; cb->args[0] = 1; info.in_skb = cb->skb; info.out_skb = skb; info.nlmsg_seq = cb->nlh->nlmsg_seq; info.nlmsg_flags = NLM_F_MULTI; list_for_each_entry(alg, &crypto_alg_list, cra_list) { err = crypto_reportstat_alg(alg, &info); if (err) goto out_err; } out: return skb->len; out_err: return err; } int crypto_dump_reportstat_done(struct netlink_callback *cb) { return 0; } MODULE_LICENSE("GPL");
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1