mirror of
				https://github.com/ossrs/srs.git
				synced 2025-03-09 15:49:59 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			849 lines
		
	
	
	
		
			25 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			849 lines
		
	
	
	
		
			25 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Copyright 2017-2021 The OpenSSL Project Authors. All Rights Reserved.
 | |
|  *
 | |
|  * Licensed under the OpenSSL license (the "License").  You may not use
 | |
|  * this file except in compliance with the License.  You can obtain a copy
 | |
|  * in the file LICENSE in the source distribution or at
 | |
|  * https://www.openssl.org/source/license.html
 | |
|  */
 | |
| 
 | |
| #include "e_os.h"
 | |
| #include <string.h>
 | |
| #include <sys/types.h>
 | |
| #include <sys/stat.h>
 | |
| #include <fcntl.h>
 | |
| #include <sys/ioctl.h>
 | |
| #include <unistd.h>
 | |
| #include <assert.h>
 | |
| 
 | |
| #include <openssl/evp.h>
 | |
| #include <openssl/err.h>
 | |
| #include <openssl/engine.h>
 | |
| #include <openssl/objects.h>
 | |
| #include <crypto/cryptodev.h>
 | |
| 
 | |
| #include "crypto/engine.h"
 | |
| 
 | |
| /* #define ENGINE_DEVCRYPTO_DEBUG */
 | |
| 
 | |
| #if CRYPTO_ALGORITHM_MIN < CRYPTO_ALGORITHM_MAX
 | |
| # define CHECK_BSD_STYLE_MACROS
 | |
| #endif
 | |
| 
 | |
| /*
 | |
|  * ONE global file descriptor for all sessions.  This allows operations
 | |
|  * such as digest session data copying (see digest_copy()), but is also
 | |
|  * saner...  why re-open /dev/crypto for every session?
 | |
|  */
 | |
| static int cfd;
 | |
| 
 | |
| static int clean_devcrypto_session(struct session_op *sess) {
 | |
|     if (ioctl(cfd, CIOCFSESSION, &sess->ses) < 0) {
 | |
|         SYSerr(SYS_F_IOCTL, errno);
 | |
|         return 0;
 | |
|     }
 | |
|     memset(sess, 0, sizeof(struct session_op));
 | |
|     return 1;
 | |
| }
 | |
| 
 | |
| /******************************************************************************
 | |
|  *
 | |
|  * Ciphers
 | |
|  *
 | |
|  * Because they all do the same basic operation, we have only one set of
 | |
|  * method functions for them all to share, and a mapping table between
 | |
|  * NIDs and cryptodev IDs, with all the necessary size data.
 | |
|  *
 | |
|  *****/
 | |
| 
 | |
| struct cipher_ctx {
 | |
|     struct session_op sess;
 | |
|     int op;                      /* COP_ENCRYPT or COP_DECRYPT */
 | |
|     unsigned long mode;          /* EVP_CIPH_*_MODE */
 | |
| 
 | |
|     /* to handle ctr mode being a stream cipher */
 | |
|     unsigned char partial[EVP_MAX_BLOCK_LENGTH];
 | |
|     unsigned int blocksize, num;
 | |
| };
 | |
| 
 | |
| static const struct cipher_data_st {
 | |
|     int nid;
 | |
|     int blocksize;
 | |
|     int keylen;
 | |
|     int ivlen;
 | |
|     int flags;
 | |
|     int devcryptoid;
 | |
| } cipher_data[] = {
 | |
| #ifndef OPENSSL_NO_DES
 | |
|     { NID_des_cbc, 8, 8, 8, EVP_CIPH_CBC_MODE, CRYPTO_DES_CBC },
 | |
|     { NID_des_ede3_cbc, 8, 24, 8, EVP_CIPH_CBC_MODE, CRYPTO_3DES_CBC },
 | |
| #endif
 | |
| #ifndef OPENSSL_NO_BF
 | |
|     { NID_bf_cbc, 8, 16, 8, EVP_CIPH_CBC_MODE, CRYPTO_BLF_CBC },
 | |
| #endif
 | |
| #ifndef OPENSSL_NO_CAST
 | |
|     { NID_cast5_cbc, 8, 16, 8, EVP_CIPH_CBC_MODE, CRYPTO_CAST_CBC },
 | |
| #endif
 | |
|     { NID_aes_128_cbc, 16, 128 / 8, 16, EVP_CIPH_CBC_MODE, CRYPTO_AES_CBC },
 | |
|     { NID_aes_192_cbc, 16, 192 / 8, 16, EVP_CIPH_CBC_MODE, CRYPTO_AES_CBC },
 | |
|     { NID_aes_256_cbc, 16, 256 / 8, 16, EVP_CIPH_CBC_MODE, CRYPTO_AES_CBC },
 | |
| #ifndef OPENSSL_NO_RC4
 | |
|     { NID_rc4, 1, 16, 0, EVP_CIPH_STREAM_CIPHER, CRYPTO_ARC4 },
 | |
| #endif
 | |
| #if !defined(CHECK_BSD_STYLE_MACROS) || defined(CRYPTO_AES_CTR)
 | |
|     { NID_aes_128_ctr, 16, 128 / 8, 16, EVP_CIPH_CTR_MODE, CRYPTO_AES_CTR },
 | |
|     { NID_aes_192_ctr, 16, 192 / 8, 16, EVP_CIPH_CTR_MODE, CRYPTO_AES_CTR },
 | |
|     { NID_aes_256_ctr, 16, 256 / 8, 16, EVP_CIPH_CTR_MODE, CRYPTO_AES_CTR },
 | |
| #endif
 | |
| #if 0                            /* Not yet supported */
 | |
|     { NID_aes_128_xts, 16, 128 / 8 * 2, 16, EVP_CIPH_XTS_MODE, CRYPTO_AES_XTS },
 | |
|     { NID_aes_256_xts, 16, 256 / 8 * 2, 16, EVP_CIPH_XTS_MODE, CRYPTO_AES_XTS },
 | |
| #endif
 | |
| #if !defined(CHECK_BSD_STYLE_MACROS) || defined(CRYPTO_AES_ECB)
 | |
|     { NID_aes_128_ecb, 16, 128 / 8, 0, EVP_CIPH_ECB_MODE, CRYPTO_AES_ECB },
 | |
|     { NID_aes_192_ecb, 16, 192 / 8, 0, EVP_CIPH_ECB_MODE, CRYPTO_AES_ECB },
 | |
|     { NID_aes_256_ecb, 16, 256 / 8, 0, EVP_CIPH_ECB_MODE, CRYPTO_AES_ECB },
 | |
| #endif
 | |
| #if 0                            /* Not yet supported */
 | |
|     { NID_aes_128_gcm, 16, 128 / 8, 16, EVP_CIPH_GCM_MODE, CRYPTO_AES_GCM },
 | |
|     { NID_aes_192_gcm, 16, 192 / 8, 16, EVP_CIPH_GCM_MODE, CRYPTO_AES_GCM },
 | |
|     { NID_aes_256_gcm, 16, 256 / 8, 16, EVP_CIPH_GCM_MODE, CRYPTO_AES_GCM },
 | |
| #endif
 | |
| #ifndef OPENSSL_NO_CAMELLIA
 | |
|     { NID_camellia_128_cbc, 16, 128 / 8, 16, EVP_CIPH_CBC_MODE,
 | |
|       CRYPTO_CAMELLIA_CBC },
 | |
|     { NID_camellia_192_cbc, 16, 192 / 8, 16, EVP_CIPH_CBC_MODE,
 | |
|       CRYPTO_CAMELLIA_CBC },
 | |
|     { NID_camellia_256_cbc, 16, 256 / 8, 16, EVP_CIPH_CBC_MODE,
 | |
|       CRYPTO_CAMELLIA_CBC },
 | |
| #endif
 | |
| };
 | |
| 
 | |
| static size_t get_cipher_data_index(int nid)
 | |
| {
 | |
|     size_t i;
 | |
| 
 | |
|     for (i = 0; i < OSSL_NELEM(cipher_data); i++)
 | |
|         if (nid == cipher_data[i].nid)
 | |
|             return i;
 | |
| 
 | |
|     /*
 | |
|      * Code further down must make sure that only NIDs in the table above
 | |
|      * are used.  If any other NID reaches this function, there's a grave
 | |
|      * coding error further down.
 | |
|      */
 | |
|     assert("Code that never should be reached" == NULL);
 | |
|     return -1;
 | |
| }
 | |
| 
 | |
| static const struct cipher_data_st *get_cipher_data(int nid)
 | |
| {
 | |
|     return &cipher_data[get_cipher_data_index(nid)];
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Following are the three necessary functions to map OpenSSL functionality
 | |
|  * with cryptodev.
 | |
|  */
 | |
| 
 | |
| static int cipher_init(EVP_CIPHER_CTX *ctx, const unsigned char *key,
 | |
|                        const unsigned char *iv, int enc)
 | |
| {
 | |
|     struct cipher_ctx *cipher_ctx =
 | |
|         (struct cipher_ctx *)EVP_CIPHER_CTX_get_cipher_data(ctx);
 | |
|     const struct cipher_data_st *cipher_d =
 | |
|         get_cipher_data(EVP_CIPHER_CTX_nid(ctx));
 | |
| 
 | |
|     /* cleanup a previous session */
 | |
|     if (cipher_ctx->sess.ses != 0 &&
 | |
|         clean_devcrypto_session(&cipher_ctx->sess) == 0)
 | |
|         return 0;
 | |
| 
 | |
|     cipher_ctx->sess.cipher = cipher_d->devcryptoid;
 | |
|     cipher_ctx->sess.keylen = cipher_d->keylen;
 | |
|     cipher_ctx->sess.key = (void *)key;
 | |
|     cipher_ctx->op = enc ? COP_ENCRYPT : COP_DECRYPT;
 | |
|     cipher_ctx->mode = cipher_d->flags & EVP_CIPH_MODE;
 | |
|     cipher_ctx->blocksize = cipher_d->blocksize;
 | |
|     if (ioctl(cfd, CIOCGSESSION, &cipher_ctx->sess) < 0) {
 | |
|         SYSerr(SYS_F_IOCTL, errno);
 | |
|         return 0;
 | |
|     }
 | |
| 
 | |
|     return 1;
 | |
| }
 | |
| 
 | |
| static int cipher_do_cipher(EVP_CIPHER_CTX *ctx, unsigned char *out,
 | |
|                             const unsigned char *in, size_t inl)
 | |
| {
 | |
|     struct cipher_ctx *cipher_ctx =
 | |
|         (struct cipher_ctx *)EVP_CIPHER_CTX_get_cipher_data(ctx);
 | |
|     struct crypt_op cryp;
 | |
|     unsigned char *iv = EVP_CIPHER_CTX_iv_noconst(ctx);
 | |
| #if !defined(COP_FLAG_WRITE_IV)
 | |
|     unsigned char saved_iv[EVP_MAX_IV_LENGTH];
 | |
|     const unsigned char *ivptr;
 | |
|     size_t nblocks, ivlen;
 | |
| #endif
 | |
| 
 | |
|     memset(&cryp, 0, sizeof(cryp));
 | |
|     cryp.ses = cipher_ctx->sess.ses;
 | |
|     cryp.len = inl;
 | |
|     cryp.src = (void *)in;
 | |
|     cryp.dst = (void *)out;
 | |
|     cryp.iv = (void *)iv;
 | |
|     cryp.op = cipher_ctx->op;
 | |
| #if !defined(COP_FLAG_WRITE_IV)
 | |
|     cryp.flags = 0;
 | |
| 
 | |
|     ivlen = EVP_CIPHER_CTX_iv_length(ctx);
 | |
|     if (ivlen > 0)
 | |
|         switch (cipher_ctx->mode) {
 | |
|         case EVP_CIPH_CBC_MODE:
 | |
|             assert(inl >= ivlen);
 | |
|             if (!EVP_CIPHER_CTX_encrypting(ctx)) {
 | |
|                 ivptr = in + inl - ivlen;
 | |
|                 memcpy(saved_iv, ivptr, ivlen);
 | |
|             }
 | |
|             break;
 | |
| 
 | |
|         case EVP_CIPH_CTR_MODE:
 | |
|             break;
 | |
| 
 | |
|         default: /* should not happen */
 | |
|             return 0;
 | |
|         }
 | |
| #else
 | |
|     cryp.flags = COP_FLAG_WRITE_IV;
 | |
| #endif
 | |
| 
 | |
|     if (ioctl(cfd, CIOCCRYPT, &cryp) < 0) {
 | |
|         SYSerr(SYS_F_IOCTL, errno);
 | |
|         return 0;
 | |
|     }
 | |
| 
 | |
| #if !defined(COP_FLAG_WRITE_IV)
 | |
|     if (ivlen > 0)
 | |
|         switch (cipher_ctx->mode) {
 | |
|         case EVP_CIPH_CBC_MODE:
 | |
|             assert(inl >= ivlen);
 | |
|             if (EVP_CIPHER_CTX_encrypting(ctx))
 | |
|                 ivptr = out + inl - ivlen;
 | |
|             else
 | |
|                 ivptr = saved_iv;
 | |
| 
 | |
|             memcpy(iv, ivptr, ivlen);
 | |
|             break;
 | |
| 
 | |
|         case EVP_CIPH_CTR_MODE:
 | |
|             nblocks = (inl + cipher_ctx->blocksize - 1)
 | |
|                       / cipher_ctx->blocksize;
 | |
|             do {
 | |
|                 ivlen--;
 | |
|                 nblocks += iv[ivlen];
 | |
|                 iv[ivlen] = (uint8_t) nblocks;
 | |
|                 nblocks >>= 8;
 | |
|             } while (ivlen);
 | |
|             break;
 | |
| 
 | |
|         default: /* should not happen */
 | |
|             return 0;
 | |
|         }
 | |
| #endif
 | |
| 
 | |
|     return 1;
 | |
| }
 | |
| 
 | |
| static int ctr_do_cipher(EVP_CIPHER_CTX *ctx, unsigned char *out,
 | |
|                          const unsigned char *in, size_t inl)
 | |
| {
 | |
|     struct cipher_ctx *cipher_ctx =
 | |
|         (struct cipher_ctx *)EVP_CIPHER_CTX_get_cipher_data(ctx);
 | |
|     size_t nblocks, len;
 | |
| 
 | |
|     /* initial partial block */
 | |
|     while (cipher_ctx->num && inl) {
 | |
|         (*out++) = *(in++) ^ cipher_ctx->partial[cipher_ctx->num];
 | |
|         --inl;
 | |
|         cipher_ctx->num = (cipher_ctx->num + 1) % cipher_ctx->blocksize;
 | |
|     }
 | |
| 
 | |
|     /* full blocks */
 | |
|     if (inl > (unsigned int) cipher_ctx->blocksize) {
 | |
|         nblocks = inl/cipher_ctx->blocksize;
 | |
|         len = nblocks * cipher_ctx->blocksize;
 | |
|         if (cipher_do_cipher(ctx, out, in, len) < 1)
 | |
|             return 0;
 | |
|         inl -= len;
 | |
|         out += len;
 | |
|         in += len;
 | |
|     }
 | |
| 
 | |
|     /* final partial block */
 | |
|     if (inl) {
 | |
|         memset(cipher_ctx->partial, 0, cipher_ctx->blocksize);
 | |
|         if (cipher_do_cipher(ctx, cipher_ctx->partial, cipher_ctx->partial,
 | |
|             cipher_ctx->blocksize) < 1)
 | |
|             return 0;
 | |
|         while (inl--) {
 | |
|             out[cipher_ctx->num] = in[cipher_ctx->num]
 | |
|                                    ^ cipher_ctx->partial[cipher_ctx->num];
 | |
|             cipher_ctx->num++;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     return 1;
 | |
| }
 | |
| 
 | |
| static int cipher_ctrl(EVP_CIPHER_CTX *ctx, int type, int p1, void* p2)
 | |
| {
 | |
|     struct cipher_ctx *cipher_ctx =
 | |
|         (struct cipher_ctx *)EVP_CIPHER_CTX_get_cipher_data(ctx);
 | |
|     EVP_CIPHER_CTX *to_ctx = (EVP_CIPHER_CTX *)p2;
 | |
|     struct cipher_ctx *to_cipher_ctx;
 | |
| 
 | |
|     switch (type) {
 | |
|     case EVP_CTRL_COPY:
 | |
|         if (cipher_ctx == NULL)
 | |
|             return 1;
 | |
|         /* when copying the context, a new session needs to be initialized */
 | |
|         to_cipher_ctx =
 | |
|             (struct cipher_ctx *)EVP_CIPHER_CTX_get_cipher_data(to_ctx);
 | |
|         memset(&to_cipher_ctx->sess, 0, sizeof(to_cipher_ctx->sess));
 | |
|         return cipher_init(to_ctx, cipher_ctx->sess.key, EVP_CIPHER_CTX_iv(ctx),
 | |
|                            (cipher_ctx->op == COP_ENCRYPT));
 | |
| 
 | |
|     case EVP_CTRL_INIT:
 | |
|         memset(&cipher_ctx->sess, 0, sizeof(cipher_ctx->sess));
 | |
|         return 1;
 | |
| 
 | |
|     default:
 | |
|         break;
 | |
|     }
 | |
| 
 | |
|     return -1;
 | |
| }
 | |
| 
 | |
| static int cipher_cleanup(EVP_CIPHER_CTX *ctx)
 | |
| {
 | |
|     struct cipher_ctx *cipher_ctx =
 | |
|         (struct cipher_ctx *)EVP_CIPHER_CTX_get_cipher_data(ctx);
 | |
| 
 | |
|     return clean_devcrypto_session(&cipher_ctx->sess);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Keep a table of known nids and associated methods.
 | |
|  * Note that known_cipher_nids[] isn't necessarily indexed the same way as
 | |
|  * cipher_data[] above, which known_cipher_methods[] is.
 | |
|  */
 | |
| static int known_cipher_nids[OSSL_NELEM(cipher_data)];
 | |
| static int known_cipher_nids_amount = -1; /* -1 indicates not yet initialised */
 | |
| static EVP_CIPHER *known_cipher_methods[OSSL_NELEM(cipher_data)] = { NULL, };
 | |
| 
 | |
| static void prepare_cipher_methods(void)
 | |
| {
 | |
|     size_t i;
 | |
|     struct session_op sess;
 | |
|     unsigned long cipher_mode;
 | |
| 
 | |
|     memset(&sess, 0, sizeof(sess));
 | |
|     sess.key = (void *)"01234567890123456789012345678901234567890123456789";
 | |
| 
 | |
|     for (i = 0, known_cipher_nids_amount = 0;
 | |
|          i < OSSL_NELEM(cipher_data); i++) {
 | |
| 
 | |
|         /*
 | |
|          * Check that the algo is really availably by trying to open and close
 | |
|          * a session.
 | |
|          */
 | |
|         sess.cipher = cipher_data[i].devcryptoid;
 | |
|         sess.keylen = cipher_data[i].keylen;
 | |
|         if (ioctl(cfd, CIOCGSESSION, &sess) < 0
 | |
|             || ioctl(cfd, CIOCFSESSION, &sess.ses) < 0)
 | |
|             continue;
 | |
| 
 | |
|         cipher_mode = cipher_data[i].flags & EVP_CIPH_MODE;
 | |
| 
 | |
|         if ((known_cipher_methods[i] =
 | |
|                  EVP_CIPHER_meth_new(cipher_data[i].nid,
 | |
|                                      cipher_mode == EVP_CIPH_CTR_MODE ? 1 :
 | |
|                                                     cipher_data[i].blocksize,
 | |
|                                      cipher_data[i].keylen)) == NULL
 | |
|             || !EVP_CIPHER_meth_set_iv_length(known_cipher_methods[i],
 | |
|                                               cipher_data[i].ivlen)
 | |
|             || !EVP_CIPHER_meth_set_flags(known_cipher_methods[i],
 | |
|                                           cipher_data[i].flags
 | |
|                                           | EVP_CIPH_CUSTOM_COPY
 | |
|                                           | EVP_CIPH_CTRL_INIT
 | |
|                                           | EVP_CIPH_FLAG_DEFAULT_ASN1)
 | |
|             || !EVP_CIPHER_meth_set_init(known_cipher_methods[i], cipher_init)
 | |
|             || !EVP_CIPHER_meth_set_do_cipher(known_cipher_methods[i],
 | |
|                                      cipher_mode == EVP_CIPH_CTR_MODE ?
 | |
|                                               ctr_do_cipher :
 | |
|                                               cipher_do_cipher)
 | |
|             || !EVP_CIPHER_meth_set_ctrl(known_cipher_methods[i], cipher_ctrl)
 | |
|             || !EVP_CIPHER_meth_set_cleanup(known_cipher_methods[i],
 | |
|                                             cipher_cleanup)
 | |
|             || !EVP_CIPHER_meth_set_impl_ctx_size(known_cipher_methods[i],
 | |
|                                                   sizeof(struct cipher_ctx))) {
 | |
|             EVP_CIPHER_meth_free(known_cipher_methods[i]);
 | |
|             known_cipher_methods[i] = NULL;
 | |
|         } else {
 | |
|             known_cipher_nids[known_cipher_nids_amount++] =
 | |
|                 cipher_data[i].nid;
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| static const EVP_CIPHER *get_cipher_method(int nid)
 | |
| {
 | |
|     size_t i = get_cipher_data_index(nid);
 | |
| 
 | |
|     if (i == (size_t)-1)
 | |
|         return NULL;
 | |
|     return known_cipher_methods[i];
 | |
| }
 | |
| 
 | |
| static int get_cipher_nids(const int **nids)
 | |
| {
 | |
|     *nids = known_cipher_nids;
 | |
|     return known_cipher_nids_amount;
 | |
| }
 | |
| 
 | |
| static void destroy_cipher_method(int nid)
 | |
| {
 | |
|     size_t i = get_cipher_data_index(nid);
 | |
| 
 | |
|     EVP_CIPHER_meth_free(known_cipher_methods[i]);
 | |
|     known_cipher_methods[i] = NULL;
 | |
| }
 | |
| 
 | |
| static void destroy_all_cipher_methods(void)
 | |
| {
 | |
|     size_t i;
 | |
| 
 | |
|     for (i = 0; i < OSSL_NELEM(cipher_data); i++)
 | |
|         destroy_cipher_method(cipher_data[i].nid);
 | |
| }
 | |
| 
 | |
| static int devcrypto_ciphers(ENGINE *e, const EVP_CIPHER **cipher,
 | |
|                              const int **nids, int nid)
 | |
| {
 | |
|     if (cipher == NULL)
 | |
|         return get_cipher_nids(nids);
 | |
| 
 | |
|     *cipher = get_cipher_method(nid);
 | |
| 
 | |
|     return *cipher != NULL;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * We only support digests if the cryptodev implementation supports multiple
 | |
|  * data updates and session copying.  Otherwise, we would be forced to maintain
 | |
|  * a cache, which is perilous if there's a lot of data coming in (if someone
 | |
|  * wants to checksum an OpenSSL tarball, for example).
 | |
|  */
 | |
| #if defined(CIOCCPHASH) && defined(COP_FLAG_UPDATE) && defined(COP_FLAG_FINAL)
 | |
| #define IMPLEMENT_DIGEST
 | |
| 
 | |
| /******************************************************************************
 | |
|  *
 | |
|  * Digests
 | |
|  *
 | |
|  * Because they all do the same basic operation, we have only one set of
 | |
|  * method functions for them all to share, and a mapping table between
 | |
|  * NIDs and cryptodev IDs, with all the necessary size data.
 | |
|  *
 | |
|  *****/
 | |
| 
 | |
| struct digest_ctx {
 | |
|     struct session_op sess;
 | |
|     /* This signals that the init function was called, not that it succeeded. */
 | |
|     int init_called;
 | |
| };
 | |
| 
 | |
| static const struct digest_data_st {
 | |
|     int nid;
 | |
|     int blocksize;
 | |
|     int digestlen;
 | |
|     int devcryptoid;
 | |
| } digest_data[] = {
 | |
| #ifndef OPENSSL_NO_MD5
 | |
|     { NID_md5, /* MD5_CBLOCK */ 64, 16, CRYPTO_MD5 },
 | |
| #endif
 | |
|     { NID_sha1, SHA_CBLOCK, 20, CRYPTO_SHA1 },
 | |
| #ifndef OPENSSL_NO_RMD160
 | |
| # if !defined(CHECK_BSD_STYLE_MACROS) || defined(CRYPTO_RIPEMD160)
 | |
|     { NID_ripemd160, /* RIPEMD160_CBLOCK */ 64, 20, CRYPTO_RIPEMD160 },
 | |
| # endif
 | |
| #endif
 | |
| #if !defined(CHECK_BSD_STYLE_MACROS) || defined(CRYPTO_SHA2_224)
 | |
|     { NID_sha224, SHA256_CBLOCK, 224 / 8, CRYPTO_SHA2_224 },
 | |
| #endif
 | |
| #if !defined(CHECK_BSD_STYLE_MACROS) || defined(CRYPTO_SHA2_256)
 | |
|     { NID_sha256, SHA256_CBLOCK, 256 / 8, CRYPTO_SHA2_256 },
 | |
| #endif
 | |
| #if !defined(CHECK_BSD_STYLE_MACROS) || defined(CRYPTO_SHA2_384)
 | |
|     { NID_sha384, SHA512_CBLOCK, 384 / 8, CRYPTO_SHA2_384 },
 | |
| #endif
 | |
| #if !defined(CHECK_BSD_STYLE_MACROS) || defined(CRYPTO_SHA2_512)
 | |
|     { NID_sha512, SHA512_CBLOCK, 512 / 8, CRYPTO_SHA2_512 },
 | |
| #endif
 | |
| };
 | |
| 
 | |
| static size_t get_digest_data_index(int nid)
 | |
| {
 | |
|     size_t i;
 | |
| 
 | |
|     for (i = 0; i < OSSL_NELEM(digest_data); i++)
 | |
|         if (nid == digest_data[i].nid)
 | |
|             return i;
 | |
| 
 | |
|     /*
 | |
|      * Code further down must make sure that only NIDs in the table above
 | |
|      * are used.  If any other NID reaches this function, there's a grave
 | |
|      * coding error further down.
 | |
|      */
 | |
|     assert("Code that never should be reached" == NULL);
 | |
|     return -1;
 | |
| }
 | |
| 
 | |
| static const struct digest_data_st *get_digest_data(int nid)
 | |
| {
 | |
|     return &digest_data[get_digest_data_index(nid)];
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Following are the four necessary functions to map OpenSSL functionality
 | |
|  * with cryptodev.
 | |
|  */
 | |
| 
 | |
| static int digest_init(EVP_MD_CTX *ctx)
 | |
| {
 | |
|     struct digest_ctx *digest_ctx =
 | |
|         (struct digest_ctx *)EVP_MD_CTX_md_data(ctx);
 | |
|     const struct digest_data_st *digest_d =
 | |
|         get_digest_data(EVP_MD_CTX_type(ctx));
 | |
| 
 | |
|     digest_ctx->init_called = 1;
 | |
| 
 | |
|     memset(&digest_ctx->sess, 0, sizeof(digest_ctx->sess));
 | |
|     digest_ctx->sess.mac = digest_d->devcryptoid;
 | |
|     if (ioctl(cfd, CIOCGSESSION, &digest_ctx->sess) < 0) {
 | |
|         SYSerr(SYS_F_IOCTL, errno);
 | |
|         return 0;
 | |
|     }
 | |
| 
 | |
|     return 1;
 | |
| }
 | |
| 
 | |
| static int digest_op(struct digest_ctx *ctx, const void *src, size_t srclen,
 | |
|                      void *res, unsigned int flags)
 | |
| {
 | |
|     struct crypt_op cryp;
 | |
| 
 | |
|     memset(&cryp, 0, sizeof(cryp));
 | |
|     cryp.ses = ctx->sess.ses;
 | |
|     cryp.len = srclen;
 | |
|     cryp.src = (void *)src;
 | |
|     cryp.dst = NULL;
 | |
|     cryp.mac = res;
 | |
|     cryp.flags = flags;
 | |
|     return ioctl(cfd, CIOCCRYPT, &cryp);
 | |
| }
 | |
| 
 | |
| static int digest_update(EVP_MD_CTX *ctx, const void *data, size_t count)
 | |
| {
 | |
|     struct digest_ctx *digest_ctx =
 | |
|         (struct digest_ctx *)EVP_MD_CTX_md_data(ctx);
 | |
| 
 | |
|     if (count == 0)
 | |
|         return 1;
 | |
| 
 | |
|     if (digest_ctx == NULL)
 | |
|         return 0;
 | |
| 
 | |
|     if (digest_op(digest_ctx, data, count, NULL, COP_FLAG_UPDATE) < 0) {
 | |
|         SYSerr(SYS_F_IOCTL, errno);
 | |
|         return 0;
 | |
|     }
 | |
| 
 | |
|     return 1;
 | |
| }
 | |
| 
 | |
| static int digest_final(EVP_MD_CTX *ctx, unsigned char *md)
 | |
| {
 | |
|     struct digest_ctx *digest_ctx =
 | |
|         (struct digest_ctx *)EVP_MD_CTX_md_data(ctx);
 | |
| 
 | |
|     if (md == NULL || digest_ctx == NULL)
 | |
|         return 0;
 | |
|     if (digest_op(digest_ctx, NULL, 0, md, COP_FLAG_FINAL) < 0) {
 | |
|         SYSerr(SYS_F_IOCTL, errno);
 | |
|         return 0;
 | |
|     }
 | |
| 
 | |
|     return 1;
 | |
| }
 | |
| 
 | |
| static int digest_copy(EVP_MD_CTX *to, const EVP_MD_CTX *from)
 | |
| {
 | |
|     struct digest_ctx *digest_from =
 | |
|         (struct digest_ctx *)EVP_MD_CTX_md_data(from);
 | |
|     struct digest_ctx *digest_to =
 | |
|         (struct digest_ctx *)EVP_MD_CTX_md_data(to);
 | |
|     struct cphash_op cphash;
 | |
| 
 | |
|     if (digest_from == NULL || digest_from->init_called != 1)
 | |
|         return 1;
 | |
| 
 | |
|     if (!digest_init(to)) {
 | |
|         SYSerr(SYS_F_IOCTL, errno);
 | |
|         return 0;
 | |
|     }
 | |
| 
 | |
|     cphash.src_ses = digest_from->sess.ses;
 | |
|     cphash.dst_ses = digest_to->sess.ses;
 | |
|     if (ioctl(cfd, CIOCCPHASH, &cphash) < 0) {
 | |
|         SYSerr(SYS_F_IOCTL, errno);
 | |
|         return 0;
 | |
|     }
 | |
|     return 1;
 | |
| }
 | |
| 
 | |
| static int digest_cleanup(EVP_MD_CTX *ctx)
 | |
| {
 | |
|     struct digest_ctx *digest_ctx =
 | |
|         (struct digest_ctx *)EVP_MD_CTX_md_data(ctx);
 | |
| 
 | |
|     if (digest_ctx == NULL)
 | |
|         return 1;
 | |
| 
 | |
|     return clean_devcrypto_session(&digest_ctx->sess);
 | |
| }
 | |
| 
 | |
| static int devcrypto_test_digest(size_t digest_data_index)
 | |
| {
 | |
|     struct session_op sess1, sess2;
 | |
|     struct cphash_op cphash;
 | |
|     int ret=0;
 | |
| 
 | |
|     memset(&sess1, 0, sizeof(sess1));
 | |
|     memset(&sess2, 0, sizeof(sess2));
 | |
|     sess1.mac = digest_data[digest_data_index].devcryptoid;
 | |
|     if (ioctl(cfd, CIOCGSESSION, &sess1) < 0)
 | |
|         return 0;
 | |
|     /* Make sure the driver is capable of hash state copy */
 | |
|     sess2.mac = sess1.mac;
 | |
|     if (ioctl(cfd, CIOCGSESSION, &sess2) >= 0) {
 | |
|         cphash.src_ses = sess1.ses;
 | |
|         cphash.dst_ses = sess2.ses;
 | |
|         if (ioctl(cfd, CIOCCPHASH, &cphash) >= 0)
 | |
|             ret = 1;
 | |
|         ioctl(cfd, CIOCFSESSION, &sess2.ses);
 | |
|     }
 | |
|     ioctl(cfd, CIOCFSESSION, &sess1.ses);
 | |
|     return ret;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Keep a table of known nids and associated methods.
 | |
|  * Note that known_digest_nids[] isn't necessarily indexed the same way as
 | |
|  * digest_data[] above, which known_digest_methods[] is.
 | |
|  */
 | |
| static int known_digest_nids[OSSL_NELEM(digest_data)];
 | |
| static int known_digest_nids_amount = -1; /* -1 indicates not yet initialised */
 | |
| static EVP_MD *known_digest_methods[OSSL_NELEM(digest_data)] = { NULL, };
 | |
| 
 | |
| static void prepare_digest_methods(void)
 | |
| {
 | |
|     size_t i;
 | |
| 
 | |
|     for (i = 0, known_digest_nids_amount = 0; i < OSSL_NELEM(digest_data);
 | |
|          i++) {
 | |
| 
 | |
|         /*
 | |
|          * Check that the algo is usable
 | |
|          */
 | |
|         if (!devcrypto_test_digest(i))
 | |
|             continue;
 | |
| 
 | |
|         if ((known_digest_methods[i] = EVP_MD_meth_new(digest_data[i].nid,
 | |
|                                                        NID_undef)) == NULL
 | |
|             || !EVP_MD_meth_set_input_blocksize(known_digest_methods[i],
 | |
|                                                 digest_data[i].blocksize)
 | |
|             || !EVP_MD_meth_set_result_size(known_digest_methods[i],
 | |
|                                             digest_data[i].digestlen)
 | |
|             || !EVP_MD_meth_set_init(known_digest_methods[i], digest_init)
 | |
|             || !EVP_MD_meth_set_update(known_digest_methods[i], digest_update)
 | |
|             || !EVP_MD_meth_set_final(known_digest_methods[i], digest_final)
 | |
|             || !EVP_MD_meth_set_copy(known_digest_methods[i], digest_copy)
 | |
|             || !EVP_MD_meth_set_cleanup(known_digest_methods[i], digest_cleanup)
 | |
|             || !EVP_MD_meth_set_app_datasize(known_digest_methods[i],
 | |
|                                              sizeof(struct digest_ctx))) {
 | |
|             EVP_MD_meth_free(known_digest_methods[i]);
 | |
|             known_digest_methods[i] = NULL;
 | |
|         } else {
 | |
|             known_digest_nids[known_digest_nids_amount++] = digest_data[i].nid;
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| static const EVP_MD *get_digest_method(int nid)
 | |
| {
 | |
|     size_t i = get_digest_data_index(nid);
 | |
| 
 | |
|     if (i == (size_t)-1)
 | |
|         return NULL;
 | |
|     return known_digest_methods[i];
 | |
| }
 | |
| 
 | |
| static int get_digest_nids(const int **nids)
 | |
| {
 | |
|     *nids = known_digest_nids;
 | |
|     return known_digest_nids_amount;
 | |
| }
 | |
| 
 | |
| static void destroy_digest_method(int nid)
 | |
| {
 | |
|     size_t i = get_digest_data_index(nid);
 | |
| 
 | |
|     EVP_MD_meth_free(known_digest_methods[i]);
 | |
|     known_digest_methods[i] = NULL;
 | |
| }
 | |
| 
 | |
| static void destroy_all_digest_methods(void)
 | |
| {
 | |
|     size_t i;
 | |
| 
 | |
|     for (i = 0; i < OSSL_NELEM(digest_data); i++)
 | |
|         destroy_digest_method(digest_data[i].nid);
 | |
| }
 | |
| 
 | |
| static int devcrypto_digests(ENGINE *e, const EVP_MD **digest,
 | |
|                              const int **nids, int nid)
 | |
| {
 | |
|     if (digest == NULL)
 | |
|         return get_digest_nids(nids);
 | |
| 
 | |
|     *digest = get_digest_method(nid);
 | |
| 
 | |
|     return *digest != NULL;
 | |
| }
 | |
| 
 | |
| #endif
 | |
| 
 | |
| /******************************************************************************
 | |
|  *
 | |
|  * LOAD / UNLOAD
 | |
|  *
 | |
|  *****/
 | |
| 
 | |
| static int devcrypto_unload(ENGINE *e)
 | |
| {
 | |
|     destroy_all_cipher_methods();
 | |
| #ifdef IMPLEMENT_DIGEST
 | |
|     destroy_all_digest_methods();
 | |
| #endif
 | |
| 
 | |
|     close(cfd);
 | |
| 
 | |
|     return 1;
 | |
| }
 | |
| /*
 | |
|  * This engine is always built into libcrypto, so it doesn't offer any
 | |
|  * ability to be dynamically loadable.
 | |
|  */
 | |
| void engine_load_devcrypto_int()
 | |
| {
 | |
|     ENGINE *e = NULL;
 | |
|     int fd;
 | |
| 
 | |
|     if ((fd = open("/dev/crypto", O_RDWR, 0)) < 0) {
 | |
| #ifndef ENGINE_DEVCRYPTO_DEBUG
 | |
|         if (errno != ENOENT)
 | |
| #endif
 | |
|             fprintf(stderr, "Could not open /dev/crypto: %s\n", strerror(errno));
 | |
|         return;
 | |
|     }
 | |
| 
 | |
| #ifdef CRIOGET
 | |
|     if (ioctl(fd, CRIOGET, &cfd) < 0) {
 | |
|         fprintf(stderr, "Could not create crypto fd: %s\n", strerror(errno));
 | |
|         close(fd);
 | |
|         cfd = -1;
 | |
|         return;
 | |
|     }
 | |
|     close(fd);
 | |
| #else
 | |
|     cfd = fd;
 | |
| #endif
 | |
| 
 | |
|     if ((e = ENGINE_new()) == NULL
 | |
|         || !ENGINE_set_destroy_function(e, devcrypto_unload)) {
 | |
|         ENGINE_free(e);
 | |
|         /*
 | |
|          * We know that devcrypto_unload() won't be called when one of the
 | |
|          * above two calls have failed, so we close cfd explicitly here to
 | |
|          * avoid leaking resources.
 | |
|          */
 | |
|         close(cfd);
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     prepare_cipher_methods();
 | |
| #ifdef IMPLEMENT_DIGEST
 | |
|     prepare_digest_methods();
 | |
| #endif
 | |
| 
 | |
|     if (!ENGINE_set_id(e, "devcrypto")
 | |
|         || !ENGINE_set_name(e, "/dev/crypto engine")
 | |
| 
 | |
| /*
 | |
|  * Asymmetric ciphers aren't well supported with /dev/crypto.  Among the BSD
 | |
|  * implementations, it seems to only exist in FreeBSD, and regarding the
 | |
|  * parameters in its crypt_kop, the manual crypto(4) has this to say:
 | |
|  *
 | |
|  *    The semantics of these arguments are currently undocumented.
 | |
|  *
 | |
|  * Reading through the FreeBSD source code doesn't give much more than
 | |
|  * their CRK_MOD_EXP implementation for ubsec.
 | |
|  *
 | |
|  * It doesn't look much better with cryptodev-linux.  They have the crypt_kop
 | |
|  * structure as well as the command (CRK_*) in cryptodev.h, but no support
 | |
|  * seems to be implemented at all for the moment.
 | |
|  *
 | |
|  * At the time of writing, it seems impossible to write proper support for
 | |
|  * FreeBSD's asym features without some very deep knowledge and access to
 | |
|  * specific kernel modules.
 | |
|  *
 | |
|  * /Richard Levitte, 2017-05-11
 | |
|  */
 | |
| #if 0
 | |
| # ifndef OPENSSL_NO_RSA
 | |
|         || !ENGINE_set_RSA(e, devcrypto_rsa)
 | |
| # endif
 | |
| # ifndef OPENSSL_NO_DSA
 | |
|         || !ENGINE_set_DSA(e, devcrypto_dsa)
 | |
| # endif
 | |
| # ifndef OPENSSL_NO_DH
 | |
|         || !ENGINE_set_DH(e, devcrypto_dh)
 | |
| # endif
 | |
| # ifndef OPENSSL_NO_EC
 | |
|         || !ENGINE_set_EC(e, devcrypto_ec)
 | |
| # endif
 | |
| #endif
 | |
|         || !ENGINE_set_ciphers(e, devcrypto_ciphers)
 | |
| #ifdef IMPLEMENT_DIGEST
 | |
|         || !ENGINE_set_digests(e, devcrypto_digests)
 | |
| #endif
 | |
|         ) {
 | |
|         ENGINE_free(e);
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     ENGINE_add(e);
 | |
|     ENGINE_free(e);          /* Loose our local reference */
 | |
|     ERR_clear_error();
 | |
| }
 |