/* SPDX-License-Identifier: GPL-2.0 * * Copyright (C) 2018 Marvell International Ltd. */ #ifndef __OCTEONTX_BCH_REGS_H__ #define __OCTEONTX_BCH_REGS_H__ #define BCH_NR_VF 1 union bch_cmd { u64 u[4]; struct fields { struct { u64 size:12; u64 reserved_12_31:20; u64 ecc_level:4; u64 reserved_36_61:26; u64 ecc_gen:2; } cword; struct { u64 ptr:49; u64 reserved_49_55:7; u64 nc:1; u64 fw:1; u64 reserved_58_63:6; } oword; struct { u64 ptr:49; u64 reserved_49_55:7; u64 nc:1; u64 reserved_57_63:7; } iword; struct { u64 ptr:49; u64 reserved_49_63:15; } rword; } s; }; enum ecc_gen { eg_correct, eg_copy, eg_gen, eg_copy3, }; /** Response from BCH instruction */ union bch_resp { u16 u16; struct { u16 num_errors:7; /** Number of errors in block */ u16 zero:6; /** Always zero, ignore */ u16 erased:1; /** Block is erased */ u16 uncorrectable:1;/** too many bits flipped */ u16 done:1; /** Block is done */ } s; }; union bch_vqx_ctl { u64 u; struct { u64 reserved_0:1; u64 cmd_be:1; u64 max_read:4; u64 reserved_6_15:10; u64 erase_disable:1; u64 one_cmd:1; u64 early_term:4; u64 reserved_22_63:42; } s; }; union bch_vqx_cmd_buf { u64 u; struct { u64 reserved_0_32:33; u64 size:13; u64 dfb:1; u64 ldwb:1; u64 reserved_48_63:16; } s; }; /* keep queue state indexed, even though just one supported here, * for later generalization to similarly-shaped queues on other Cavium devices */ enum { QID_BCH, QID_MAX }; struct bch_q { struct udevice *dev; int index; u16 max_depth; u16 pool_size_m1; u64 *base_vaddr; dma_addr_t base_paddr; }; extern struct bch_q octeontx_bch_q[QID_MAX]; /* with one dma-mapped area, virt<->phys conversions by +/- (vaddr-paddr) */ static inline dma_addr_t qphys(int qid, void *v) { struct bch_q *q = &octeontx_bch_q[qid]; int off = (u8 *)v - (u8 *)q->base_vaddr; return q->base_paddr + off; } #define octeontx_ptr_to_phys(v) qphys(QID_BCH, (v)) static inline void *qvirt(int qid, dma_addr_t p) { struct bch_q *q = &octeontx_bch_q[qid]; int off = p - q->base_paddr; return q->base_vaddr + off; } #define octeontx_phys_to_ptr(p) qvirt(QID_BCH, (p)) /* plenty for interleaved r/w on two planes with 16k page, ecc_size 1k */ /* QDEPTH >= 16, as successive chunks must align on 128-byte boundaries */ #define QDEPTH 256 /* u64s in a command queue chunk, incl next-pointer */ #define NQS 1 /* linked chunks in the chain */ /** * Write an arbitrary number of command words to a command queue. * This is a generic function; the fixed number of command word * functions yield higher performance. * * Could merge with crypto version for FPA use on cn83xx */ static inline int octeontx_cmd_queue_write(int queue_id, bool use_locking, int cmd_count, const u64 *cmds) { int ret = 0; u64 *cmd_ptr; struct bch_q *qptr = &octeontx_bch_q[queue_id]; if (unlikely(cmd_count < 1 || cmd_count > 32)) return -EINVAL; if (unlikely(!cmds)) return -EINVAL; cmd_ptr = qptr->base_vaddr; while (cmd_count > 0) { int slot = qptr->index % (QDEPTH * NQS); if (slot % QDEPTH != QDEPTH - 1) { cmd_ptr[slot] = *cmds++; cmd_count--; } qptr->index++; } __iowmb(); /* flush commands before ringing bell */ return ret; } #endif /* __OCTEONTX_BCH_REGS_H__ */