1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /* 3 * Asynchronous Compression operations 4 * 5 * Copyright (c) 2016, Intel Corporation 6 * Authors: Weigang Li <weigang.li@intel.com> 7 * Giovanni Cabiddu <giovanni.cabiddu@intel.com> 8 */ 9 10 #include <crypto/internal/acompress.h> 11 #include <linux/cryptouser.h> 12 #include <linux/errno.h> 13 #include <linux/kernel.h> 14 #include <linux/module.h> 15 #include <linux/seq_file.h> 16 #include <linux/slab.h> 17 #include <linux/string.h> 18 #include <net/netlink.h> 19 20 #include "compress.h" 21 22 struct crypto_scomp; 23 24 static const struct crypto_type crypto_acomp_type; 25 26 static inline struct acomp_alg *__crypto_acomp_alg(struct crypto_alg *alg) 27 { 28 return container_of(alg, struct acomp_alg, calg.base); 29 } 30 31 static inline struct acomp_alg *crypto_acomp_alg(struct crypto_acomp *tfm) 32 { 33 return __crypto_acomp_alg(crypto_acomp_tfm(tfm)->__crt_alg); 34 } 35 36 static int __maybe_unused crypto_acomp_report( 37 struct sk_buff *skb, struct crypto_alg *alg) 38 { 39 struct crypto_report_acomp racomp; 40 41 memset(&racomp, 0, sizeof(racomp)); 42 43 strscpy(racomp.type, "acomp", sizeof(racomp.type)); 44 45 return nla_put(skb, CRYPTOCFGA_REPORT_ACOMP, sizeof(racomp), &racomp); 46 } 47 48 static void crypto_acomp_show(struct seq_file *m, struct crypto_alg *alg) 49 __maybe_unused; 50 51 static void crypto_acomp_show(struct seq_file *m, struct crypto_alg *alg) 52 { 53 seq_puts(m, "type : acomp\n"); 54 } 55 56 static void crypto_acomp_exit_tfm(struct crypto_tfm *tfm) 57 { 58 struct crypto_acomp *acomp = __crypto_acomp_tfm(tfm); 59 struct acomp_alg *alg = crypto_acomp_alg(acomp); 60 61 alg->exit(acomp); 62 } 63 64 static int crypto_acomp_init_tfm(struct crypto_tfm *tfm) 65 { 66 struct crypto_acomp *acomp = __crypto_acomp_tfm(tfm); 67 struct acomp_alg *alg = crypto_acomp_alg(acomp); 68 69 if (tfm->__crt_alg->cra_type != &crypto_acomp_type) 70 return crypto_init_scomp_ops_async(tfm); 71 72 acomp->compress = alg->compress; 73 acomp->decompress = alg->decompress; 74 acomp->dst_free = alg->dst_free; 75 acomp->reqsize = alg->reqsize; 76 77 if (alg->exit) 78 acomp->base.exit = crypto_acomp_exit_tfm; 79 80 if (alg->init) 81 return alg->init(acomp); 82 83 return 0; 84 } 85 86 static unsigned int crypto_acomp_extsize(struct crypto_alg *alg) 87 { 88 int extsize = crypto_alg_extsize(alg); 89 90 if (alg->cra_type != &crypto_acomp_type) 91 extsize += sizeof(struct crypto_scomp *); 92 93 return extsize; 94 } 95 96 static inline int __crypto_acomp_report_stat(struct sk_buff *skb, 97 struct crypto_alg *alg) 98 { 99 struct comp_alg_common *calg = __crypto_comp_alg_common(alg); 100 struct crypto_istat_compress *istat = comp_get_stat(calg); 101 struct crypto_stat_compress racomp; 102 103 memset(&racomp, 0, sizeof(racomp)); 104 105 strscpy(racomp.type, "acomp", sizeof(racomp.type)); 106 racomp.stat_compress_cnt = atomic64_read(&istat->compress_cnt); 107 racomp.stat_compress_tlen = atomic64_read(&istat->compress_tlen); 108 racomp.stat_decompress_cnt = atomic64_read(&istat->decompress_cnt); 109 racomp.stat_decompress_tlen = atomic64_read(&istat->decompress_tlen); 110 racomp.stat_err_cnt = atomic64_read(&istat->err_cnt); 111 112 return nla_put(skb, CRYPTOCFGA_STAT_ACOMP, sizeof(racomp), &racomp); 113 } 114 115 #ifdef CONFIG_CRYPTO_STATS 116 int crypto_acomp_report_stat(struct sk_buff *skb, struct crypto_alg *alg) 117 { 118 return __crypto_acomp_report_stat(skb, alg); 119 } 120 #endif 121 122 static const struct crypto_type crypto_acomp_type = { 123 .extsize = crypto_acomp_extsize, 124 .init_tfm = crypto_acomp_init_tfm, 125 #ifdef CONFIG_PROC_FS 126 .show = crypto_acomp_show, 127 #endif 128 #if IS_ENABLED(CONFIG_CRYPTO_USER) 129 .report = crypto_acomp_report, 130 #endif 131 #ifdef CONFIG_CRYPTO_STATS 132 .report_stat = crypto_acomp_report_stat, 133 #endif 134 .maskclear = ~CRYPTO_ALG_TYPE_MASK, 135 .maskset = CRYPTO_ALG_TYPE_ACOMPRESS_MASK, 136 .type = CRYPTO_ALG_TYPE_ACOMPRESS, 137 .tfmsize = offsetof(struct crypto_acomp, base), 138 }; 139 140 struct crypto_acomp *crypto_alloc_acomp(const char *alg_name, u32 type, 141 u32 mask) 142 { 143 return crypto_alloc_tfm(alg_name, &crypto_acomp_type, type, mask); 144 } 145 EXPORT_SYMBOL_GPL(crypto_alloc_acomp); 146 147 struct crypto_acomp *crypto_alloc_acomp_node(const char *alg_name, u32 type, 148 u32 mask, int node) 149 { 150 return crypto_alloc_tfm_node(alg_name, &crypto_acomp_type, type, mask, 151 node); 152 } 153 EXPORT_SYMBOL_GPL(crypto_alloc_acomp_node); 154 155 struct acomp_req *acomp_request_alloc(struct crypto_acomp *acomp) 156 { 157 struct crypto_tfm *tfm = crypto_acomp_tfm(acomp); 158 struct acomp_req *req; 159 160 req = __acomp_request_alloc(acomp); 161 if (req && (tfm->__crt_alg->cra_type != &crypto_acomp_type)) 162 return crypto_acomp_scomp_alloc_ctx(req); 163 164 return req; 165 } 166 EXPORT_SYMBOL_GPL(acomp_request_alloc); 167 168 void acomp_request_free(struct acomp_req *req) 169 { 170 struct crypto_acomp *acomp = crypto_acomp_reqtfm(req); 171 struct crypto_tfm *tfm = crypto_acomp_tfm(acomp); 172 173 if (tfm->__crt_alg->cra_type != &crypto_acomp_type) 174 crypto_acomp_scomp_free_ctx(req); 175 176 if (req->flags & CRYPTO_ACOMP_ALLOC_OUTPUT) { 177 acomp->dst_free(req->dst); 178 req->dst = NULL; 179 } 180 181 __acomp_request_free(req); 182 } 183 EXPORT_SYMBOL_GPL(acomp_request_free); 184 185 void comp_prepare_alg(struct comp_alg_common *alg) 186 { 187 struct crypto_istat_compress *istat = comp_get_stat(alg); 188 struct crypto_alg *base = &alg->base; 189 190 base->cra_flags &= ~CRYPTO_ALG_TYPE_MASK; 191 192 if (IS_ENABLED(CONFIG_CRYPTO_STATS)) 193 memset(istat, 0, sizeof(*istat)); 194 } 195 196 int crypto_register_acomp(struct acomp_alg *alg) 197 { 198 struct crypto_alg *base = &alg->calg.base; 199 200 comp_prepare_alg(&alg->calg); 201 202 base->cra_type = &crypto_acomp_type; 203 base->cra_flags |= CRYPTO_ALG_TYPE_ACOMPRESS; 204 205 return crypto_register_alg(base); 206 } 207 EXPORT_SYMBOL_GPL(crypto_register_acomp); 208 209 void crypto_unregister_acomp(struct acomp_alg *alg) 210 { 211 crypto_unregister_alg(&alg->base); 212 } 213 EXPORT_SYMBOL_GPL(crypto_unregister_acomp); 214 215 int crypto_register_acomps(struct acomp_alg *algs, int count) 216 { 217 int i, ret; 218 219 for (i = 0; i < count; i++) { 220 ret = crypto_register_acomp(&algs[i]); 221 if (ret) 222 goto err; 223 } 224 225 return 0; 226 227 err: 228 for (--i; i >= 0; --i) 229 crypto_unregister_acomp(&algs[i]); 230 231 return ret; 232 } 233 EXPORT_SYMBOL_GPL(crypto_register_acomps); 234 235 void crypto_unregister_acomps(struct acomp_alg *algs, int count) 236 { 237 int i; 238 239 for (i = count - 1; i >= 0; --i) 240 crypto_unregister_acomp(&algs[i]); 241 } 242 EXPORT_SYMBOL_GPL(crypto_unregister_acomps); 243 244 MODULE_LICENSE("GPL"); 245 MODULE_DESCRIPTION("Asynchronous compression type"); 246