mirror of
https://github.com/torvalds/linux.git
synced 2026-04-18 06:44:00 -04:00
- sec_register_to_crypto() and sec_unregister_from_crypto() have been removed, the function declarations have not been removed. Remove them. - hisi_qm_start_qp and hisi_qm_stop_qp are called internally by the QM. Therefore, the EXPORT_SYMBOL_GPL declaration of these non-public interfaces is deleted. Signed-off-by: Chenghai Huang <huangchenghai2@huawei.com> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
290 lines
6.0 KiB
C
290 lines
6.0 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/* Copyright (c) 2019 HiSilicon Limited. */
|
|
|
|
#ifndef __HISI_SEC_V2_H
|
|
#define __HISI_SEC_V2_H
|
|
|
|
#include <linux/hisi_acc_qm.h>
|
|
#include "sec_crypto.h"
|
|
|
|
#define SEC_PBUF_SZ 512
|
|
#define SEC_MAX_MAC_LEN 64
|
|
#define SEC_IV_SIZE 24
|
|
#define SEC_SGE_NR_NUM 4
|
|
#define SEC_SGL_ALIGN_SIZE 64
|
|
|
|
/* Algorithm resource per hardware SEC queue */
|
|
struct sec_alg_res {
|
|
u8 *pbuf;
|
|
dma_addr_t pbuf_dma;
|
|
u8 *c_ivin;
|
|
dma_addr_t c_ivin_dma;
|
|
u8 *a_ivin;
|
|
dma_addr_t a_ivin_dma;
|
|
u8 *out_mac;
|
|
dma_addr_t out_mac_dma;
|
|
u16 depth;
|
|
};
|
|
|
|
struct sec_hw_sge {
|
|
dma_addr_t buf;
|
|
void *page_ctrl;
|
|
__le32 len;
|
|
__le32 pad;
|
|
__le32 pad0;
|
|
__le32 pad1;
|
|
};
|
|
|
|
struct sec_hw_sgl {
|
|
dma_addr_t next_dma;
|
|
__le16 entry_sum_in_chain;
|
|
__le16 entry_sum_in_sgl;
|
|
__le16 entry_length_in_sgl;
|
|
__le16 pad0;
|
|
__le64 pad1[5];
|
|
struct sec_hw_sgl *next;
|
|
struct sec_hw_sge sge_entries[SEC_SGE_NR_NUM];
|
|
} __aligned(SEC_SGL_ALIGN_SIZE);
|
|
|
|
struct sec_src_dst_buf {
|
|
struct sec_hw_sgl in;
|
|
struct sec_hw_sgl out;
|
|
};
|
|
|
|
struct sec_request_buf {
|
|
union {
|
|
struct sec_src_dst_buf data_buf;
|
|
__u8 pbuf[SEC_PBUF_SZ];
|
|
};
|
|
dma_addr_t in_dma;
|
|
dma_addr_t out_dma;
|
|
};
|
|
|
|
/* Cipher request of SEC private */
|
|
struct sec_cipher_req {
|
|
struct hisi_acc_hw_sgl *c_out;
|
|
dma_addr_t c_out_dma;
|
|
u8 *c_ivin;
|
|
dma_addr_t c_ivin_dma;
|
|
struct skcipher_request *sk_req;
|
|
u32 c_len;
|
|
bool encrypt;
|
|
__u8 c_ivin_buf[SEC_IV_SIZE];
|
|
};
|
|
|
|
struct sec_aead_req {
|
|
u8 *out_mac;
|
|
dma_addr_t out_mac_dma;
|
|
u8 *a_ivin;
|
|
dma_addr_t a_ivin_dma;
|
|
struct aead_request *aead_req;
|
|
__u8 a_ivin_buf[SEC_IV_SIZE];
|
|
__u8 out_mac_buf[SEC_MAX_MAC_LEN];
|
|
};
|
|
|
|
/* SEC request of Crypto */
|
|
struct sec_req {
|
|
union {
|
|
struct sec_sqe sec_sqe;
|
|
struct sec_sqe3 sec_sqe3;
|
|
};
|
|
struct sec_ctx *ctx;
|
|
struct sec_qp_ctx *qp_ctx;
|
|
|
|
/**
|
|
* Common parameter of the SEC request.
|
|
*/
|
|
struct hisi_acc_hw_sgl *in;
|
|
dma_addr_t in_dma;
|
|
struct sec_cipher_req c_req;
|
|
struct sec_aead_req aead_req;
|
|
struct crypto_async_request *base;
|
|
|
|
int err_type;
|
|
int req_id;
|
|
u32 flag;
|
|
|
|
bool use_pbuf;
|
|
|
|
struct list_head list;
|
|
struct sec_request_buf buf;
|
|
};
|
|
|
|
/**
|
|
* struct sec_req_op - Operations for SEC request
|
|
* @buf_map: DMA map the SGL buffers of the request
|
|
* @buf_unmap: DMA unmap the SGL buffers of the request
|
|
* @bd_fill: Fill the SEC queue BD
|
|
* @bd_send: Send the SEC BD into the hardware queue
|
|
* @callback: Call back for the request
|
|
* @process: Main processing logic of Skcipher
|
|
*/
|
|
struct sec_req_op {
|
|
int (*buf_map)(struct sec_ctx *ctx, struct sec_req *req);
|
|
void (*buf_unmap)(struct sec_ctx *ctx, struct sec_req *req);
|
|
void (*do_transfer)(struct sec_ctx *ctx, struct sec_req *req);
|
|
int (*bd_fill)(struct sec_ctx *ctx, struct sec_req *req);
|
|
int (*bd_send)(struct sec_ctx *ctx, struct sec_req *req);
|
|
void (*callback)(struct sec_ctx *ctx, struct sec_req *req, int err);
|
|
int (*process)(struct sec_ctx *ctx, struct sec_req *req);
|
|
};
|
|
|
|
/* SEC auth context */
|
|
struct sec_auth_ctx {
|
|
dma_addr_t a_key_dma;
|
|
u8 *a_key;
|
|
u8 a_key_len;
|
|
u8 a_alg;
|
|
struct crypto_shash *hash_tfm;
|
|
struct crypto_aead *fallback_aead_tfm;
|
|
};
|
|
|
|
/* SEC cipher context which cipher's relatives */
|
|
struct sec_cipher_ctx {
|
|
u8 *c_key;
|
|
dma_addr_t c_key_dma;
|
|
sector_t iv_offset;
|
|
u32 c_gran_size;
|
|
u32 ivsize;
|
|
u8 c_mode;
|
|
u8 c_alg;
|
|
u8 c_key_len;
|
|
|
|
/* add software support */
|
|
bool fallback;
|
|
struct crypto_sync_skcipher *fbtfm;
|
|
};
|
|
|
|
/* SEC queue context which defines queue's relatives */
|
|
struct sec_qp_ctx {
|
|
struct hisi_qp *qp;
|
|
struct sec_req **req_list;
|
|
struct idr req_idr;
|
|
struct sec_alg_res *res;
|
|
struct sec_ctx *ctx;
|
|
spinlock_t req_lock;
|
|
spinlock_t id_lock;
|
|
struct hisi_acc_sgl_pool *c_in_pool;
|
|
struct hisi_acc_sgl_pool *c_out_pool;
|
|
u16 send_head;
|
|
};
|
|
|
|
enum sec_alg_type {
|
|
SEC_SKCIPHER,
|
|
SEC_AEAD
|
|
};
|
|
|
|
/* SEC Crypto TFM context which defines queue and cipher .etc relatives */
|
|
struct sec_ctx {
|
|
struct sec_qp_ctx *qp_ctx;
|
|
struct sec_dev *sec;
|
|
const struct sec_req_op *req_op;
|
|
struct hisi_qp **qps;
|
|
|
|
/* Half queues for encipher, and half for decipher */
|
|
u32 hlf_q_num;
|
|
|
|
/* Current cyclic index to select a queue for encipher */
|
|
atomic_t enc_qcyclic;
|
|
|
|
/* Current cyclic index to select a queue for decipher */
|
|
atomic_t dec_qcyclic;
|
|
|
|
enum sec_alg_type alg_type;
|
|
bool pbuf_supported;
|
|
struct sec_cipher_ctx c_ctx;
|
|
struct sec_auth_ctx a_ctx;
|
|
u8 type_supported;
|
|
struct device *dev;
|
|
};
|
|
|
|
|
|
enum sec_debug_file_index {
|
|
SEC_CLEAR_ENABLE,
|
|
SEC_DEBUG_FILE_NUM,
|
|
};
|
|
|
|
struct sec_debug_file {
|
|
enum sec_debug_file_index index;
|
|
spinlock_t lock;
|
|
struct hisi_qm *qm;
|
|
};
|
|
|
|
struct sec_dfx {
|
|
atomic64_t send_cnt;
|
|
atomic64_t recv_cnt;
|
|
atomic64_t send_busy_cnt;
|
|
atomic64_t recv_busy_cnt;
|
|
atomic64_t err_bd_cnt;
|
|
atomic64_t invalid_req_cnt;
|
|
atomic64_t done_flag_cnt;
|
|
};
|
|
|
|
struct sec_debug {
|
|
struct sec_dfx dfx;
|
|
struct sec_debug_file files[SEC_DEBUG_FILE_NUM];
|
|
};
|
|
|
|
struct sec_dev {
|
|
struct hisi_qm qm;
|
|
struct sec_debug debug;
|
|
u32 ctx_q_num;
|
|
bool iommu_used;
|
|
};
|
|
|
|
enum sec_cap_type {
|
|
SEC_QM_NFE_MASK_CAP = 0x0,
|
|
SEC_QM_RESET_MASK_CAP,
|
|
SEC_QM_OOO_SHUTDOWN_MASK_CAP,
|
|
SEC_QM_CE_MASK_CAP,
|
|
SEC_NFE_MASK_CAP,
|
|
SEC_RESET_MASK_CAP,
|
|
SEC_OOO_SHUTDOWN_MASK_CAP,
|
|
SEC_CE_MASK_CAP,
|
|
SEC_CLUSTER_NUM_CAP,
|
|
SEC_CORE_TYPE_NUM_CAP,
|
|
SEC_CORE_NUM_CAP,
|
|
SEC_CORES_PER_CLUSTER_NUM_CAP,
|
|
SEC_CORE_ENABLE_BITMAP,
|
|
SEC_DRV_ALG_BITMAP_LOW,
|
|
SEC_DRV_ALG_BITMAP_HIGH,
|
|
SEC_DEV_ALG_BITMAP_LOW,
|
|
SEC_DEV_ALG_BITMAP_HIGH,
|
|
SEC_CORE1_ALG_BITMAP_LOW,
|
|
SEC_CORE1_ALG_BITMAP_HIGH,
|
|
SEC_CORE2_ALG_BITMAP_LOW,
|
|
SEC_CORE2_ALG_BITMAP_HIGH,
|
|
SEC_CORE3_ALG_BITMAP_LOW,
|
|
SEC_CORE3_ALG_BITMAP_HIGH,
|
|
SEC_CORE4_ALG_BITMAP_LOW,
|
|
SEC_CORE4_ALG_BITMAP_HIGH,
|
|
};
|
|
|
|
enum sec_cap_table_type {
|
|
QM_RAS_NFE_TYPE = 0x0,
|
|
QM_RAS_NFE_RESET,
|
|
QM_RAS_CE_TYPE,
|
|
SEC_RAS_NFE_TYPE,
|
|
SEC_RAS_NFE_RESET,
|
|
SEC_RAS_CE_TYPE,
|
|
SEC_CORE_INFO,
|
|
SEC_CORE_EN,
|
|
SEC_DRV_ALG_BITMAP_LOW_TB,
|
|
SEC_DRV_ALG_BITMAP_HIGH_TB,
|
|
SEC_ALG_BITMAP_LOW,
|
|
SEC_ALG_BITMAP_HIGH,
|
|
SEC_CORE1_BITMAP_LOW,
|
|
SEC_CORE1_BITMAP_HIGH,
|
|
SEC_CORE2_BITMAP_LOW,
|
|
SEC_CORE2_BITMAP_HIGH,
|
|
SEC_CORE3_BITMAP_LOW,
|
|
SEC_CORE3_BITMAP_HIGH,
|
|
SEC_CORE4_BITMAP_LOW,
|
|
SEC_CORE4_BITMAP_HIGH,
|
|
};
|
|
|
|
void sec_destroy_qps(struct hisi_qp **qps, int qp_num);
|
|
struct hisi_qp **sec_create_qps(void);
|
|
u64 sec_get_alg_bitmap(struct hisi_qm *qm, u32 high, u32 low);
|
|
#endif
|