mirror of
https://github.com/torvalds/linux.git
synced 2026-04-18 06:44:00 -04:00
IMA computes hashes using the crypto_shash or crypto_ahash API. The latter is used only when ima.ahash_minsize is set on the command line, and its purpose is ostensibly to make the hash computation faster. However, going off the CPU to a crypto engine and back again is actually quite slow, especially compared with the acceleration that is built into modern CPUs and the kernel now enables by default for most algorithms. Typical performance results for SHA-256 on a modern platform can be found at https://lore.kernel.org/linux-crypto/20250615184638.GA1480@sol/ Partly for this reason, several other kernel subsystems have already dropped support for the crypto_ahash API. The other problem with crypto_ahash is that bugs are also common, not just in the underlying drivers, but also in the code using it, since it is very difficult to use correctly. Just from a quick review, here are some of the bugs I noticed in IMA's ahash code: - [Use after free] ima_alloc_atfm() isn't thread-safe and can trigger a use-after-free if multiple threads try to initialize the global ima_ahash_tfm at the same time. - [Deadlock] If only one buffer is allocated and there is an error reading from the file, then ahash_wait() is executed twice, causing a deadlock in wait_for_completion(). - [Crash or incorrect hash computed] calc_buffer_ahash_atfm() is sometimes passed stack buffers which can be vmalloc addresses, but it puts them in a scatterlist assuming they are linear addresses. This causes the hashing to be done on the wrong physical address. - [Truncation to 32-bit length] ima_alloc_pages() incorrectly assumes an loff_t value fits in an unsigned long. calc_buffer_ahash_atfm() incorrectly assumes that a loff_t value fits in an unsigned int. So, not exactly a great track record so far, even disregarding driver bugs which are an even larger problem. Fortunately, in practice it's unlikely that many users are actually setting the ima.ahash_minsize kernel command-line parameter which enables this code. However, given that this code is almost certainly no longer useful (if it ever was), let's just remove it instead of attempting to fix all these issues. Signed-off-by: Eric Biggers <ebiggers@kernel.org> Acked-by: Dmitry Kasatkin <dmitry.kasatkin@gmail.com> Signed-off-by: Mimi Zohar <zohar@linux.ibm.com>
520 lines
12 KiB
C
520 lines
12 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Copyright (C) 2005,2006,2007,2008 IBM Corporation
|
|
*
|
|
* Authors:
|
|
* Mimi Zohar <zohar@us.ibm.com>
|
|
* Kylene Hall <kjhall@us.ibm.com>
|
|
*
|
|
* File: ima_crypto.c
|
|
* Calculates md5/sha1 file hash, template hash, boot-aggreate hash
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/file.h>
|
|
#include <linux/crypto.h>
|
|
#include <linux/err.h>
|
|
#include <linux/slab.h>
|
|
#include <crypto/hash.h>
|
|
|
|
#include "ima.h"
|
|
|
|
static struct crypto_shash *ima_shash_tfm;
|
|
|
|
int ima_sha1_idx __ro_after_init;
|
|
int ima_hash_algo_idx __ro_after_init;
|
|
/*
|
|
* Additional number of slots reserved, as needed, for SHA1
|
|
* and IMA default algo.
|
|
*/
|
|
int ima_extra_slots __ro_after_init;
|
|
|
|
struct ima_algo_desc *ima_algo_array __ro_after_init;
|
|
|
|
static int __init ima_init_ima_crypto(void)
|
|
{
|
|
long rc;
|
|
|
|
ima_shash_tfm = crypto_alloc_shash(hash_algo_name[ima_hash_algo], 0, 0);
|
|
if (IS_ERR(ima_shash_tfm)) {
|
|
rc = PTR_ERR(ima_shash_tfm);
|
|
pr_err("Can not allocate %s (reason: %ld)\n",
|
|
hash_algo_name[ima_hash_algo], rc);
|
|
return rc;
|
|
}
|
|
pr_info("Allocated hash algorithm: %s\n",
|
|
hash_algo_name[ima_hash_algo]);
|
|
return 0;
|
|
}
|
|
|
|
static struct crypto_shash *ima_alloc_tfm(enum hash_algo algo)
|
|
{
|
|
struct crypto_shash *tfm = ima_shash_tfm;
|
|
int rc, i;
|
|
|
|
if (algo < 0 || algo >= HASH_ALGO__LAST)
|
|
algo = ima_hash_algo;
|
|
|
|
if (algo == ima_hash_algo)
|
|
return tfm;
|
|
|
|
for (i = 0; i < NR_BANKS(ima_tpm_chip) + ima_extra_slots; i++)
|
|
if (ima_algo_array[i].tfm && ima_algo_array[i].algo == algo)
|
|
return ima_algo_array[i].tfm;
|
|
|
|
tfm = crypto_alloc_shash(hash_algo_name[algo], 0, 0);
|
|
if (IS_ERR(tfm)) {
|
|
rc = PTR_ERR(tfm);
|
|
pr_err("Can not allocate %s (reason: %d)\n",
|
|
hash_algo_name[algo], rc);
|
|
}
|
|
return tfm;
|
|
}
|
|
|
|
int __init ima_init_crypto(void)
|
|
{
|
|
unsigned int digest_size;
|
|
enum hash_algo algo;
|
|
long rc;
|
|
int i;
|
|
|
|
rc = ima_init_ima_crypto();
|
|
if (rc)
|
|
return rc;
|
|
|
|
ima_sha1_idx = -1;
|
|
ima_hash_algo_idx = -1;
|
|
|
|
for (i = 0; i < NR_BANKS(ima_tpm_chip); i++) {
|
|
algo = ima_tpm_chip->allocated_banks[i].crypto_id;
|
|
if (algo == HASH_ALGO_SHA1)
|
|
ima_sha1_idx = i;
|
|
|
|
if (algo == ima_hash_algo)
|
|
ima_hash_algo_idx = i;
|
|
}
|
|
|
|
if (ima_sha1_idx < 0) {
|
|
ima_sha1_idx = NR_BANKS(ima_tpm_chip) + ima_extra_slots++;
|
|
if (ima_hash_algo == HASH_ALGO_SHA1)
|
|
ima_hash_algo_idx = ima_sha1_idx;
|
|
}
|
|
|
|
if (ima_hash_algo_idx < 0)
|
|
ima_hash_algo_idx = NR_BANKS(ima_tpm_chip) + ima_extra_slots++;
|
|
|
|
ima_algo_array = kzalloc_objs(*ima_algo_array,
|
|
NR_BANKS(ima_tpm_chip) + ima_extra_slots);
|
|
if (!ima_algo_array) {
|
|
rc = -ENOMEM;
|
|
goto out;
|
|
}
|
|
|
|
for (i = 0; i < NR_BANKS(ima_tpm_chip); i++) {
|
|
algo = ima_tpm_chip->allocated_banks[i].crypto_id;
|
|
digest_size = ima_tpm_chip->allocated_banks[i].digest_size;
|
|
ima_algo_array[i].algo = algo;
|
|
ima_algo_array[i].digest_size = digest_size;
|
|
|
|
/* unknown TPM algorithm */
|
|
if (algo == HASH_ALGO__LAST)
|
|
continue;
|
|
|
|
if (algo == ima_hash_algo) {
|
|
ima_algo_array[i].tfm = ima_shash_tfm;
|
|
continue;
|
|
}
|
|
|
|
ima_algo_array[i].tfm = ima_alloc_tfm(algo);
|
|
if (IS_ERR(ima_algo_array[i].tfm)) {
|
|
if (algo == HASH_ALGO_SHA1) {
|
|
rc = PTR_ERR(ima_algo_array[i].tfm);
|
|
ima_algo_array[i].tfm = NULL;
|
|
goto out_array;
|
|
}
|
|
|
|
ima_algo_array[i].tfm = NULL;
|
|
}
|
|
}
|
|
|
|
if (ima_sha1_idx >= NR_BANKS(ima_tpm_chip)) {
|
|
if (ima_hash_algo == HASH_ALGO_SHA1) {
|
|
ima_algo_array[ima_sha1_idx].tfm = ima_shash_tfm;
|
|
} else {
|
|
ima_algo_array[ima_sha1_idx].tfm =
|
|
ima_alloc_tfm(HASH_ALGO_SHA1);
|
|
if (IS_ERR(ima_algo_array[ima_sha1_idx].tfm)) {
|
|
rc = PTR_ERR(ima_algo_array[ima_sha1_idx].tfm);
|
|
goto out_array;
|
|
}
|
|
}
|
|
|
|
ima_algo_array[ima_sha1_idx].algo = HASH_ALGO_SHA1;
|
|
ima_algo_array[ima_sha1_idx].digest_size = SHA1_DIGEST_SIZE;
|
|
}
|
|
|
|
if (ima_hash_algo_idx >= NR_BANKS(ima_tpm_chip) &&
|
|
ima_hash_algo_idx != ima_sha1_idx) {
|
|
digest_size = hash_digest_size[ima_hash_algo];
|
|
ima_algo_array[ima_hash_algo_idx].tfm = ima_shash_tfm;
|
|
ima_algo_array[ima_hash_algo_idx].algo = ima_hash_algo;
|
|
ima_algo_array[ima_hash_algo_idx].digest_size = digest_size;
|
|
}
|
|
|
|
return 0;
|
|
out_array:
|
|
for (i = 0; i < NR_BANKS(ima_tpm_chip) + ima_extra_slots; i++) {
|
|
if (!ima_algo_array[i].tfm ||
|
|
ima_algo_array[i].tfm == ima_shash_tfm)
|
|
continue;
|
|
|
|
crypto_free_shash(ima_algo_array[i].tfm);
|
|
}
|
|
kfree(ima_algo_array);
|
|
out:
|
|
crypto_free_shash(ima_shash_tfm);
|
|
return rc;
|
|
}
|
|
|
|
static void ima_free_tfm(struct crypto_shash *tfm)
|
|
{
|
|
int i;
|
|
|
|
if (tfm == ima_shash_tfm)
|
|
return;
|
|
|
|
for (i = 0; i < NR_BANKS(ima_tpm_chip) + ima_extra_slots; i++)
|
|
if (ima_algo_array[i].tfm == tfm)
|
|
return;
|
|
|
|
crypto_free_shash(tfm);
|
|
}
|
|
|
|
static int ima_calc_file_hash_tfm(struct file *file,
|
|
struct ima_digest_data *hash,
|
|
struct crypto_shash *tfm)
|
|
{
|
|
loff_t i_size, offset = 0;
|
|
char *rbuf;
|
|
int rc;
|
|
SHASH_DESC_ON_STACK(shash, tfm);
|
|
|
|
shash->tfm = tfm;
|
|
|
|
hash->length = crypto_shash_digestsize(tfm);
|
|
|
|
rc = crypto_shash_init(shash);
|
|
if (rc != 0)
|
|
return rc;
|
|
|
|
i_size = i_size_read(file_inode(file));
|
|
|
|
if (i_size == 0)
|
|
goto out;
|
|
|
|
rbuf = kzalloc(PAGE_SIZE, GFP_KERNEL);
|
|
if (!rbuf)
|
|
return -ENOMEM;
|
|
|
|
while (offset < i_size) {
|
|
int rbuf_len;
|
|
|
|
rbuf_len = integrity_kernel_read(file, offset, rbuf, PAGE_SIZE);
|
|
if (rbuf_len < 0) {
|
|
rc = rbuf_len;
|
|
break;
|
|
}
|
|
if (rbuf_len == 0) { /* unexpected EOF */
|
|
rc = -EINVAL;
|
|
break;
|
|
}
|
|
offset += rbuf_len;
|
|
|
|
rc = crypto_shash_update(shash, rbuf, rbuf_len);
|
|
if (rc)
|
|
break;
|
|
}
|
|
kfree(rbuf);
|
|
out:
|
|
if (!rc)
|
|
rc = crypto_shash_final(shash, hash->digest);
|
|
return rc;
|
|
}
|
|
|
|
/*
|
|
* ima_calc_file_hash - calculate file hash
|
|
*/
|
|
int ima_calc_file_hash(struct file *file, struct ima_digest_data *hash)
|
|
{
|
|
int rc;
|
|
struct file *f = file;
|
|
bool new_file_instance = false;
|
|
struct crypto_shash *tfm;
|
|
|
|
/*
|
|
* For consistency, fail file's opened with the O_DIRECT flag on
|
|
* filesystems mounted with/without DAX option.
|
|
*/
|
|
if (file->f_flags & O_DIRECT) {
|
|
hash->length = hash_digest_size[ima_hash_algo];
|
|
hash->algo = ima_hash_algo;
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Open a new file instance in O_RDONLY if we cannot read */
|
|
if (!(file->f_mode & FMODE_READ)) {
|
|
int flags = file->f_flags & ~(O_WRONLY | O_APPEND |
|
|
O_TRUNC | O_CREAT | O_NOCTTY | O_EXCL);
|
|
flags |= O_RDONLY;
|
|
f = dentry_open(&file->f_path, flags, file->f_cred);
|
|
if (IS_ERR(f))
|
|
return PTR_ERR(f);
|
|
|
|
new_file_instance = true;
|
|
}
|
|
|
|
tfm = ima_alloc_tfm(hash->algo);
|
|
if (IS_ERR(tfm)) {
|
|
rc = PTR_ERR(tfm);
|
|
} else {
|
|
rc = ima_calc_file_hash_tfm(f, hash, tfm);
|
|
ima_free_tfm(tfm);
|
|
}
|
|
if (new_file_instance)
|
|
fput(f);
|
|
return rc;
|
|
}
|
|
|
|
/*
|
|
* Calculate the hash of template data
|
|
*/
|
|
static int ima_calc_field_array_hash_tfm(struct ima_field_data *field_data,
|
|
struct ima_template_entry *entry,
|
|
int tfm_idx)
|
|
{
|
|
SHASH_DESC_ON_STACK(shash, ima_algo_array[tfm_idx].tfm);
|
|
struct ima_template_desc *td = entry->template_desc;
|
|
int num_fields = entry->template_desc->num_fields;
|
|
int rc, i;
|
|
|
|
shash->tfm = ima_algo_array[tfm_idx].tfm;
|
|
|
|
rc = crypto_shash_init(shash);
|
|
if (rc != 0)
|
|
return rc;
|
|
|
|
for (i = 0; i < num_fields; i++) {
|
|
u8 buffer[IMA_EVENT_NAME_LEN_MAX + 1] = { 0 };
|
|
u8 *data_to_hash = field_data[i].data;
|
|
u32 datalen = field_data[i].len;
|
|
u32 datalen_to_hash = !ima_canonical_fmt ?
|
|
datalen : (__force u32)cpu_to_le32(datalen);
|
|
|
|
if (strcmp(td->name, IMA_TEMPLATE_IMA_NAME) != 0) {
|
|
rc = crypto_shash_update(shash,
|
|
(const u8 *) &datalen_to_hash,
|
|
sizeof(datalen_to_hash));
|
|
if (rc)
|
|
break;
|
|
} else if (strcmp(td->fields[i]->field_id, "n") == 0) {
|
|
memcpy(buffer, data_to_hash, datalen);
|
|
data_to_hash = buffer;
|
|
datalen = IMA_EVENT_NAME_LEN_MAX + 1;
|
|
}
|
|
rc = crypto_shash_update(shash, data_to_hash, datalen);
|
|
if (rc)
|
|
break;
|
|
}
|
|
|
|
if (!rc)
|
|
rc = crypto_shash_final(shash, entry->digests[tfm_idx].digest);
|
|
|
|
return rc;
|
|
}
|
|
|
|
int ima_calc_field_array_hash(struct ima_field_data *field_data,
|
|
struct ima_template_entry *entry)
|
|
{
|
|
u16 alg_id;
|
|
int rc, i;
|
|
|
|
rc = ima_calc_field_array_hash_tfm(field_data, entry, ima_sha1_idx);
|
|
if (rc)
|
|
return rc;
|
|
|
|
entry->digests[ima_sha1_idx].alg_id = TPM_ALG_SHA1;
|
|
|
|
for (i = 0; i < NR_BANKS(ima_tpm_chip) + ima_extra_slots; i++) {
|
|
if (i == ima_sha1_idx)
|
|
continue;
|
|
|
|
if (i < NR_BANKS(ima_tpm_chip)) {
|
|
alg_id = ima_tpm_chip->allocated_banks[i].alg_id;
|
|
entry->digests[i].alg_id = alg_id;
|
|
}
|
|
|
|
/* for unmapped TPM algorithms digest is still a padded SHA1 */
|
|
if (!ima_algo_array[i].tfm) {
|
|
memcpy(entry->digests[i].digest,
|
|
entry->digests[ima_sha1_idx].digest,
|
|
TPM_DIGEST_SIZE);
|
|
continue;
|
|
}
|
|
|
|
rc = ima_calc_field_array_hash_tfm(field_data, entry, i);
|
|
if (rc)
|
|
return rc;
|
|
}
|
|
return rc;
|
|
}
|
|
|
|
static int calc_buffer_shash_tfm(const void *buf, loff_t size,
|
|
struct ima_digest_data *hash,
|
|
struct crypto_shash *tfm)
|
|
{
|
|
SHASH_DESC_ON_STACK(shash, tfm);
|
|
unsigned int len;
|
|
int rc;
|
|
|
|
shash->tfm = tfm;
|
|
|
|
hash->length = crypto_shash_digestsize(tfm);
|
|
|
|
rc = crypto_shash_init(shash);
|
|
if (rc != 0)
|
|
return rc;
|
|
|
|
while (size) {
|
|
len = size < PAGE_SIZE ? size : PAGE_SIZE;
|
|
rc = crypto_shash_update(shash, buf, len);
|
|
if (rc)
|
|
break;
|
|
buf += len;
|
|
size -= len;
|
|
}
|
|
|
|
if (!rc)
|
|
rc = crypto_shash_final(shash, hash->digest);
|
|
return rc;
|
|
}
|
|
|
|
int ima_calc_buffer_hash(const void *buf, loff_t len,
|
|
struct ima_digest_data *hash)
|
|
{
|
|
struct crypto_shash *tfm;
|
|
int rc;
|
|
|
|
tfm = ima_alloc_tfm(hash->algo);
|
|
if (IS_ERR(tfm))
|
|
return PTR_ERR(tfm);
|
|
|
|
rc = calc_buffer_shash_tfm(buf, len, hash, tfm);
|
|
|
|
ima_free_tfm(tfm);
|
|
return rc;
|
|
}
|
|
|
|
static void ima_pcrread(u32 idx, struct tpm_digest *d)
|
|
{
|
|
if (!ima_tpm_chip)
|
|
return;
|
|
|
|
if (tpm_pcr_read(ima_tpm_chip, idx, d) != 0)
|
|
pr_err("Error Communicating to TPM chip\n");
|
|
}
|
|
|
|
/*
|
|
* The boot_aggregate is a cumulative hash over TPM registers 0 - 7. With
|
|
* TPM 1.2 the boot_aggregate was based on reading the SHA1 PCRs, but with
|
|
* TPM 2.0 hash agility, TPM chips could support multiple TPM PCR banks,
|
|
* allowing firmware to configure and enable different banks.
|
|
*
|
|
* Knowing which TPM bank is read to calculate the boot_aggregate digest
|
|
* needs to be conveyed to a verifier. For this reason, use the same
|
|
* hash algorithm for reading the TPM PCRs as for calculating the boot
|
|
* aggregate digest as stored in the measurement list.
|
|
*/
|
|
static int ima_calc_boot_aggregate_tfm(char *digest, u16 alg_id,
|
|
struct crypto_shash *tfm)
|
|
{
|
|
struct tpm_digest d = { .alg_id = alg_id, .digest = {0} };
|
|
int rc;
|
|
u32 i;
|
|
SHASH_DESC_ON_STACK(shash, tfm);
|
|
|
|
shash->tfm = tfm;
|
|
|
|
pr_devel("calculating the boot-aggregate based on TPM bank: %04x\n",
|
|
d.alg_id);
|
|
|
|
rc = crypto_shash_init(shash);
|
|
if (rc != 0)
|
|
return rc;
|
|
|
|
/* cumulative digest over TPM registers 0-7 */
|
|
for (i = TPM_PCR0; i < TPM_PCR8; i++) {
|
|
ima_pcrread(i, &d);
|
|
/* now accumulate with current aggregate */
|
|
rc = crypto_shash_update(shash, d.digest,
|
|
crypto_shash_digestsize(tfm));
|
|
if (rc != 0)
|
|
return rc;
|
|
}
|
|
/*
|
|
* Extend cumulative digest over TPM registers 8-9, which contain
|
|
* measurement for the kernel command line (reg. 8) and image (reg. 9)
|
|
* in a typical PCR allocation. Registers 8-9 are only included in
|
|
* non-SHA1 boot_aggregate digests to avoid ambiguity.
|
|
*/
|
|
if (alg_id != TPM_ALG_SHA1) {
|
|
for (i = TPM_PCR8; i < TPM_PCR10; i++) {
|
|
ima_pcrread(i, &d);
|
|
rc = crypto_shash_update(shash, d.digest,
|
|
crypto_shash_digestsize(tfm));
|
|
}
|
|
}
|
|
if (!rc)
|
|
rc = crypto_shash_final(shash, digest);
|
|
return rc;
|
|
}
|
|
|
|
int ima_calc_boot_aggregate(struct ima_digest_data *hash)
|
|
{
|
|
struct crypto_shash *tfm;
|
|
u16 crypto_id, alg_id;
|
|
int rc, i, bank_idx = -1;
|
|
|
|
for (i = 0; i < ima_tpm_chip->nr_allocated_banks; i++) {
|
|
crypto_id = ima_tpm_chip->allocated_banks[i].crypto_id;
|
|
if (crypto_id == hash->algo) {
|
|
bank_idx = i;
|
|
break;
|
|
}
|
|
|
|
if (crypto_id == HASH_ALGO_SHA256)
|
|
bank_idx = i;
|
|
|
|
if (bank_idx == -1 && crypto_id == HASH_ALGO_SHA1)
|
|
bank_idx = i;
|
|
}
|
|
|
|
if (bank_idx == -1) {
|
|
pr_err("No suitable TPM algorithm for boot aggregate\n");
|
|
return 0;
|
|
}
|
|
|
|
hash->algo = ima_tpm_chip->allocated_banks[bank_idx].crypto_id;
|
|
|
|
tfm = ima_alloc_tfm(hash->algo);
|
|
if (IS_ERR(tfm))
|
|
return PTR_ERR(tfm);
|
|
|
|
hash->length = crypto_shash_digestsize(tfm);
|
|
alg_id = ima_tpm_chip->allocated_banks[bank_idx].alg_id;
|
|
rc = ima_calc_boot_aggregate_tfm(hash->digest, alg_id, tfm);
|
|
|
|
ima_free_tfm(tfm);
|
|
|
|
return rc;
|
|
}
|