Commit b4b7e67c authored by Tadeusz Struk's avatar Tadeusz Struk Committed by Herbert Xu

crypto: qat - Intel(R) QAT ucode part of fw loader

This patch adds microcode part of the firmware loader.

v4 - splits FW loader part into two smaller patches.
Acked-by: default avatarBo Cui <bo.cui@intel.com>
Reviewed-by: default avatarBruce W. Allan <bruce.w.allan@intel.com>
Signed-off-by: default avatarKaren Xiang <karen.xiang@intel.com>
Signed-off-by: default avatarPingchaox Yang <pingchaox.yang@intel.com>
Signed-off-by: default avatarHerbert Xu <herbert@gondor.apana.org.au>
parent d370cec3
/*
This file is provided under a dual BSD/GPLv2 license. When using or
redistributing this file, you may do so under either license.
GPL LICENSE SUMMARY
Copyright(c) 2014 Intel Corporation.
This program is free software; you can redistribute it and/or modify
it under the terms of version 2 of the GNU General Public License as
published by the Free Software Foundation.
This program is distributed in the hope that it will be useful, but
WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
General Public License for more details.
Contact Information:
qat-linux@intel.com
BSD LICENSE
Copyright(c) 2014 Intel Corporation.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions
are met:
* Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
* Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in
the documentation and/or other materials provided with the
distribution.
* Neither the name of Intel Corporation nor the names of its
contributors may be used to endorse or promote products derived
from this software without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#ifndef __ICP_QAT_UCLO_H__
#define __ICP_QAT_UCLO_H__
#define ICP_QAT_AC_C_CPU_TYPE 0x00400000
#define ICP_QAT_UCLO_MAX_AE 12
#define ICP_QAT_UCLO_MAX_CTX 8
#define ICP_QAT_UCLO_MAX_UIMAGE (ICP_QAT_UCLO_MAX_AE * ICP_QAT_UCLO_MAX_CTX)
#define ICP_QAT_UCLO_MAX_USTORE 0x4000
#define ICP_QAT_UCLO_MAX_XFER_REG 128
#define ICP_QAT_UCLO_MAX_GPR_REG 128
#define ICP_QAT_UCLO_MAX_NN_REG 128
#define ICP_QAT_UCLO_MAX_LMEM_REG 1024
#define ICP_QAT_UCLO_AE_ALL_CTX 0xff
#define ICP_QAT_UOF_OBJID_LEN 8
#define ICP_QAT_UOF_FID 0xc6c2
#define ICP_QAT_UOF_MAJVER 0x4
#define ICP_QAT_UOF_MINVER 0x11
#define ICP_QAT_UOF_NN_MODE_NOTCARE 0xff
#define ICP_QAT_UOF_OBJS "UOF_OBJS"
#define ICP_QAT_UOF_STRT "UOF_STRT"
#define ICP_QAT_UOF_GTID "UOF_GTID"
#define ICP_QAT_UOF_IMAG "UOF_IMAG"
#define ICP_QAT_UOF_IMEM "UOF_IMEM"
#define ICP_QAT_UOF_MSEG "UOF_MSEG"
#define ICP_QAT_UOF_LOCAL_SCOPE 1
#define ICP_QAT_UOF_INIT_EXPR 0
#define ICP_QAT_UOF_INIT_REG 1
#define ICP_QAT_UOF_INIT_REG_CTX 2
#define ICP_QAT_UOF_INIT_EXPR_ENDIAN_SWAP 3
#define ICP_QAT_CTX_MODE(ae_mode) ((ae_mode) & 0xf)
#define ICP_QAT_NN_MODE(ae_mode) (((ae_mode) >> 0x4) & 0xf)
#define ICP_QAT_SHARED_USTORE_MODE(ae_mode) (((ae_mode) >> 0xb) & 0x1)
#define RELOADABLE_CTX_SHARED_MODE(ae_mode) (((ae_mode) >> 0xc) & 0x1)
#define ICP_QAT_LOC_MEM0_MODE(ae_mode) (((ae_mode) >> 0x8) & 0x1)
#define ICP_QAT_LOC_MEM1_MODE(ae_mode) (((ae_mode) >> 0x9) & 0x1)
enum icp_qat_uof_mem_region {
ICP_QAT_UOF_SRAM_REGION = 0x0,
ICP_QAT_UOF_LMEM_REGION = 0x3,
ICP_QAT_UOF_UMEM_REGION = 0x5
};
enum icp_qat_uof_regtype {
ICP_NO_DEST,
ICP_GPA_REL,
ICP_GPA_ABS,
ICP_GPB_REL,
ICP_GPB_ABS,
ICP_SR_REL,
ICP_SR_RD_REL,
ICP_SR_WR_REL,
ICP_SR_ABS,
ICP_SR_RD_ABS,
ICP_SR_WR_ABS,
ICP_DR_REL,
ICP_DR_RD_REL,
ICP_DR_WR_REL,
ICP_DR_ABS,
ICP_DR_RD_ABS,
ICP_DR_WR_ABS,
ICP_LMEM,
ICP_LMEM0,
ICP_LMEM1,
ICP_NEIGH_REL,
};
struct icp_qat_uclo_page {
struct icp_qat_uclo_encap_page *encap_page;
struct icp_qat_uclo_region *region;
unsigned int flags;
};
struct icp_qat_uclo_region {
struct icp_qat_uclo_page *loaded;
struct icp_qat_uclo_page *page;
};
struct icp_qat_uclo_aeslice {
struct icp_qat_uclo_region *regions;
struct icp_qat_uclo_page *page;
struct icp_qat_uclo_page *cur_page[ICP_QAT_UCLO_MAX_CTX];
struct icp_qat_uclo_encapme *encap_image;
unsigned int ctx_mask_assigned;
unsigned int new_uaddr[ICP_QAT_UCLO_MAX_CTX];
};
struct icp_qat_uclo_aedata {
unsigned int slice_num;
unsigned int eff_ustore_size;
struct icp_qat_uclo_aeslice ae_slices[ICP_QAT_UCLO_MAX_CTX];
};
struct icp_qat_uof_encap_obj {
char *beg_uof;
struct icp_qat_uof_objhdr *obj_hdr;
struct icp_qat_uof_chunkhdr *chunk_hdr;
struct icp_qat_uof_varmem_seg *var_mem_seg;
};
struct icp_qat_uclo_encap_uwblock {
unsigned int start_addr;
unsigned int words_num;
uint64_t micro_words;
};
struct icp_qat_uclo_encap_page {
unsigned int def_page;
unsigned int page_region;
unsigned int beg_addr_v;
unsigned int beg_addr_p;
unsigned int micro_words_num;
unsigned int uwblock_num;
struct icp_qat_uclo_encap_uwblock *uwblock;
};
struct icp_qat_uclo_encapme {
struct icp_qat_uof_image *img_ptr;
struct icp_qat_uclo_encap_page *page;
unsigned int ae_reg_num;
struct icp_qat_uof_ae_reg *ae_reg;
unsigned int init_regsym_num;
struct icp_qat_uof_init_regsym *init_regsym;
unsigned int sbreak_num;
struct icp_qat_uof_sbreak *sbreak;
unsigned int uwords_num;
};
struct icp_qat_uclo_init_mem_table {
unsigned int entry_num;
struct icp_qat_uof_initmem *init_mem;
};
struct icp_qat_uclo_objhdr {
char *file_buff;
unsigned int checksum;
unsigned int size;
};
struct icp_qat_uof_strtable {
unsigned int table_len;
unsigned int reserved;
uint64_t strings;
};
struct icp_qat_uclo_objhandle {
unsigned int prod_type;
unsigned int prod_rev;
struct icp_qat_uclo_objhdr *obj_hdr;
struct icp_qat_uof_encap_obj encap_uof_obj;
struct icp_qat_uof_strtable str_table;
struct icp_qat_uclo_encapme ae_uimage[ICP_QAT_UCLO_MAX_UIMAGE];
struct icp_qat_uclo_aedata ae_data[ICP_QAT_UCLO_MAX_AE];
struct icp_qat_uclo_init_mem_table init_mem_tab;
struct icp_qat_uof_batch_init *lm_init_tab[ICP_QAT_UCLO_MAX_AE];
struct icp_qat_uof_batch_init *umem_init_tab[ICP_QAT_UCLO_MAX_AE];
int uimage_num;
int uword_in_bytes;
int global_inited;
unsigned int ae_num;
unsigned int ustore_phy_size;
void *obj_buf;
uint64_t *uword_buf;
};
struct icp_qat_uof_uword_block {
unsigned int start_addr;
unsigned int words_num;
unsigned int uword_offset;
unsigned int reserved;
};
struct icp_qat_uof_filehdr {
unsigned short file_id;
unsigned short reserved1;
char min_ver;
char maj_ver;
unsigned short reserved2;
unsigned short max_chunks;
unsigned short num_chunks;
};
struct icp_qat_uof_filechunkhdr {
char chunk_id[ICP_QAT_UOF_OBJID_LEN];
unsigned int checksum;
unsigned int offset;
unsigned int size;
};
struct icp_qat_uof_objhdr {
unsigned int cpu_type;
unsigned short min_cpu_ver;
unsigned short max_cpu_ver;
short max_chunks;
short num_chunks;
unsigned int reserved1;
unsigned int reserved2;
};
struct icp_qat_uof_chunkhdr {
char chunk_id[ICP_QAT_UOF_OBJID_LEN];
unsigned int offset;
unsigned int size;
};
struct icp_qat_uof_memvar_attr {
unsigned int offset_in_byte;
unsigned int value;
};
struct icp_qat_uof_initmem {
unsigned int sym_name;
char region;
char scope;
unsigned short reserved1;
unsigned int addr;
unsigned int num_in_bytes;
unsigned int val_attr_num;
};
struct icp_qat_uof_init_regsym {
unsigned int sym_name;
char init_type;
char value_type;
char reg_type;
unsigned char ctx;
unsigned int reg_addr;
unsigned int value;
};
struct icp_qat_uof_varmem_seg {
unsigned int sram_base;
unsigned int sram_size;
unsigned int sram_alignment;
unsigned int sdram_base;
unsigned int sdram_size;
unsigned int sdram_alignment;
unsigned int sdram1_base;
unsigned int sdram1_size;
unsigned int sdram1_alignment;
unsigned int scratch_base;
unsigned int scratch_size;
unsigned int scratch_alignment;
};
struct icp_qat_uof_gtid {
char tool_id[ICP_QAT_UOF_OBJID_LEN];
int tool_ver;
unsigned int reserved1;
unsigned int reserved2;
};
struct icp_qat_uof_sbreak {
unsigned int page_num;
unsigned int virt_uaddr;
unsigned char sbreak_type;
unsigned char reg_type;
unsigned short reserved1;
unsigned int addr_offset;
unsigned int reg_addr;
};
struct icp_qat_uof_code_page {
unsigned int page_region;
unsigned int page_num;
unsigned char def_page;
unsigned char reserved2;
unsigned short reserved1;
unsigned int beg_addr_v;
unsigned int beg_addr_p;
unsigned int neigh_reg_tab_offset;
unsigned int uc_var_tab_offset;
unsigned int imp_var_tab_offset;
unsigned int imp_expr_tab_offset;
unsigned int code_area_offset;
};
struct icp_qat_uof_image {
unsigned int img_name;
unsigned int ae_assigned;
unsigned int ctx_assigned;
unsigned int cpu_type;
unsigned int entry_address;
unsigned int fill_pattern[2];
unsigned int reloadable_size;
unsigned char sensitivity;
unsigned char reserved;
unsigned short ae_mode;
unsigned short max_ver;
unsigned short min_ver;
unsigned short image_attrib;
unsigned short reserved2;
unsigned short page_region_num;
unsigned short numpages;
unsigned int reg_tab_offset;
unsigned int init_reg_sym_tab;
unsigned int sbreak_tab;
unsigned int app_metadata;
};
struct icp_qat_uof_objtable {
unsigned int entry_num;
};
struct icp_qat_uof_ae_reg {
unsigned int name;
unsigned int vis_name;
unsigned short type;
unsigned short addr;
unsigned short access_mode;
unsigned char visible;
unsigned char reserved1;
unsigned short ref_count;
unsigned short reserved2;
unsigned int xo_id;
};
struct icp_qat_uof_code_area {
unsigned int micro_words_num;
unsigned int uword_block_tab;
};
struct icp_qat_uof_batch_init {
unsigned int ae;
unsigned int addr;
unsigned int *value;
unsigned int size;
struct icp_qat_uof_batch_init *next;
};
#endif
/*
This file is provided under a dual BSD/GPLv2 license. When using or
redistributing this file, you may do so under either license.
GPL LICENSE SUMMARY
Copyright(c) 2014 Intel Corporation.
This program is free software; you can redistribute it and/or modify
it under the terms of version 2 of the GNU General Public License as
published by the Free Software Foundation.
This program is distributed in the hope that it will be useful, but
WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
General Public License for more details.
Contact Information:
qat-linux@intel.com
BSD LICENSE
Copyright(c) 2014 Intel Corporation.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions
are met:
* Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
* Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in
the documentation and/or other materials provided with the
distribution.
* Neither the name of Intel Corporation nor the names of its
contributors may be used to endorse or promote products derived
from this software without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include <linux/slab.h>
#include <linux/ctype.h>
#include <linux/kernel.h>
#include "adf_accel_devices.h"
#include "adf_common_drv.h"
#include "icp_qat_uclo.h"
#include "icp_qat_hal.h"
#include "icp_qat_fw_loader_handle.h"
#define UWORD_CPYBUF_SIZE 1024
#define INVLD_UWORD 0xffffffffffull
#define PID_MINOR_REV 0xf
#define PID_MAJOR_REV (0xf << 4)
static int qat_uclo_init_ae_data(struct icp_qat_uclo_objhandle *obj_handle,
unsigned int ae, unsigned int image_num)
{
struct icp_qat_uclo_aedata *ae_data;
struct icp_qat_uclo_encapme *encap_image;
struct icp_qat_uclo_page *page = NULL;
struct icp_qat_uclo_aeslice *ae_slice = NULL;
ae_data = &obj_handle->ae_data[ae];
encap_image = &obj_handle->ae_uimage[image_num];
ae_slice = &ae_data->ae_slices[ae_data->slice_num];
ae_slice->encap_image = encap_image;
if (encap_image->img_ptr) {
ae_slice->ctx_mask_assigned =
encap_image->img_ptr->ctx_assigned;
ae_data->eff_ustore_size = obj_handle->ustore_phy_size;
} else {
ae_slice->ctx_mask_assigned = 0;
}
ae_slice->regions = kzalloc(sizeof(*(ae_slice->regions)), GFP_KERNEL);
if (!(ae_slice->regions))
return -ENOMEM;
ae_slice->page = kzalloc(sizeof(*(ae_slice->page)), GFP_KERNEL);
if (!(ae_slice->page))
goto out_err;
page = ae_slice->page;
page->encap_page = encap_image->page;
ae_slice->page->region = ae_slice->regions;
ae_data->slice_num++;
return 0;
out_err:
kfree(ae_slice->regions);
ae_slice->regions = NULL;
return -ENOMEM;
}
static int qat_uclo_free_ae_data(struct icp_qat_uclo_aedata *ae_data)
{
unsigned int ss = 0;
if (!ae_data) {
pr_err("QAT: bad argument, ae_data is NULL\n ");
return -EINVAL;
}
for (ss = 0; ss < ae_data->slice_num; ss++) {
kfree(ae_data->ae_slices[ss].regions);
ae_data->ae_slices[ss].regions = NULL;
kfree(ae_data->ae_slices[ss].page);
ae_data->ae_slices[ss].page = NULL;
}
return 0;
}
static char *qat_uclo_get_string(struct icp_qat_uof_strtable *str_table,
unsigned int str_offset)
{
if ((!str_table->table_len) || (str_offset > str_table->table_len))
return NULL;
return (char *)(((unsigned long)(str_table->strings)) + str_offset);
}
static int qat_uclo_check_format(struct icp_qat_uof_filehdr *hdr)
{
int maj = hdr->maj_ver & 0xff;
int min = hdr->min_ver & 0xff;
if (hdr->file_id != ICP_QAT_UOF_FID) {
pr_err("QAT: Invalid header 0x%x\n", hdr->file_id);
return -EINVAL;
}
if (min != ICP_QAT_UOF_MINVER || maj != ICP_QAT_UOF_MAJVER) {
pr_err("QAT: bad uof version, major 0x%x, minor 0x%x\n",
maj, min);
return -EINVAL;
}
return 0;
}
static void qat_uclo_wr_sram_by_words(struct icp_qat_fw_loader_handle *handle,
unsigned int addr, unsigned int *val,
unsigned int num_in_bytes)
{
unsigned int outval;
unsigned char *ptr = (unsigned char *)val;
while (num_in_bytes) {
memcpy(&outval, ptr, 4);
SRAM_WRITE(handle, addr, outval);
num_in_bytes -= 4;
ptr += 4;
addr += 4;
}
}
static void qat_uclo_wr_umem_by_words(struct icp_qat_fw_loader_handle *handle,
unsigned char ae, unsigned int addr,
unsigned int *val,
unsigned int num_in_bytes)
{
unsigned int outval;
unsigned char *ptr = (unsigned char *)val;
addr >>= 0x2; /* convert to uword address */
while (num_in_bytes) {
memcpy(&outval, ptr, 4);
qat_hal_wr_umem(handle, ae, addr++, 1, &outval);
num_in_bytes -= 4;
ptr += 4;
}
return;
}
static void qat_uclo_batch_wr_umem(struct icp_qat_fw_loader_handle *handle,
unsigned char ae,
struct icp_qat_uof_batch_init
*umem_init_header)
{
struct icp_qat_uof_batch_init *umem_init;
if (!umem_init_header)
return;
umem_init = umem_init_header->next;
while (umem_init) {
unsigned int addr, *value, size;
ae = umem_init->ae;
addr = umem_init->addr;
value = umem_init->value;
size = umem_init->size;
qat_uclo_wr_umem_by_words(handle, ae, addr, value, size);
umem_init = umem_init->next;
}
}
static void
qat_uclo_cleanup_batch_init_list(struct icp_qat_fw_loader_handle *handle,
struct icp_qat_uof_batch_init **base)
{
struct icp_qat_uof_batch_init *umem_init;
umem_init = *base;
while (umem_init) {
struct icp_qat_uof_batch_init *pre;
pre = umem_init;
umem_init = umem_init->next;
kfree(pre);
}
*base = NULL;
}
static int qat_uclo_parse_num(char *str, unsigned int *num)
{
char buf[16];
unsigned long ae = 0;
int i;
memset(buf, '\0', 16);
strncpy(buf, str, 15);
for (i = 0; i < 16; i++) {
if (!isdigit(buf[i])) {
buf[i] = '\0';
break;
}
}
if ((kstrtoul(buf, 10, &ae)))
return -EFAULT;
*num = (unsigned int)ae;
return 0;
}
static int qat_uclo_fetch_initmem_ae(struct icp_qat_fw_loader_handle *handle,
struct icp_qat_uof_initmem *init_mem,
unsigned int size_range, unsigned int *ae)
{
struct icp_qat_uclo_objhandle *obj_handle = handle->obj_handle;
char *str;
if ((init_mem->addr + init_mem->num_in_bytes) > (size_range << 0x2)) {
pr_err("QAT: initmem is out of range");
return -EINVAL;
}
if (init_mem->scope != ICP_QAT_UOF_LOCAL_SCOPE) {
pr_err("QAT: Memory scope for init_mem error\n");
return -EINVAL;
}
str = qat_uclo_get_string(&(obj_handle->str_table), init_mem->sym_name);
if (!str) {
pr_err("QAT: AE name assigned in uof init table is NULL\n");
return -EINVAL;
}
if (qat_uclo_parse_num(str, ae)) {
pr_err("QAT: Parse num for AE number failed\n");
return -EINVAL;
}
if (!test_bit(*ae, (unsigned long *)&(handle->hal_handle->ae_mask))) {
pr_err("QAT: ae %d to be init is fused off\n", *ae);
return -EINVAL;
}
if (*ae >= ICP_QAT_UCLO_MAX_AE) {
pr_err("QAT: ae %d out of range\n", *ae);
return -EINVAL;
}
return 0;
}
static int qat_uclo_create_batch_init_list(struct icp_qat_fw_loader_handle
*handle, struct icp_qat_uof_initmem
*init_mem, unsigned int ae,
struct icp_qat_uof_batch_init
**init_tab_base)
{
struct icp_qat_uof_batch_init *init_header, *tail;
struct icp_qat_uof_batch_init *mem_init, *tail_old;
struct icp_qat_uof_memvar_attr *mem_val_attr;
unsigned int i, flag = 0;
mem_val_attr =
(struct icp_qat_uof_memvar_attr *)((unsigned long)init_mem +
sizeof(struct icp_qat_uof_initmem));
init_header = *init_tab_base;
if (!init_header) {
init_header = kzalloc(sizeof(*init_header), GFP_KERNEL);
if (!init_header)
return -ENOMEM;
init_header->size = 1;
*init_tab_base = init_header;
flag = 1;
}
tail_old = init_header;
while (tail_old->next)
tail_old = tail_old->next;
tail = tail_old;
for (i = 0; i < init_mem->val_attr_num; i++) {
mem_init = kzalloc(sizeof(*mem_init), GFP_KERNEL);
if (!mem_init)
goto out_err;
mem_init->ae = ae;
mem_init->addr = init_mem->addr + mem_val_attr->offset_in_byte;
mem_init->value = &mem_val_attr->value;
mem_init->size = 4;
mem_init->next = NULL;
tail->next = mem_init;
tail = mem_init;
init_header->size += qat_hal_get_ins_num();
mem_val_attr++;
}
return 0;
out_err:
while (tail_old) {
mem_init = tail_old->next;
kfree(tail_old);
tail_old = mem_init;
}
if (flag)
kfree(*init_tab_base);
return -ENOMEM;
}
static int qat_uclo_init_lmem_seg(struct icp_qat_fw_loader_handle *handle,
struct icp_qat_uof_initmem *init_mem)
{
struct icp_qat_uclo_objhandle *obj_handle = handle->obj_handle;
unsigned int ae;
if (qat_uclo_fetch_initmem_ae(handle, init_mem,
ICP_QAT_UCLO_MAX_LMEM_REG, &ae))
return -EINVAL;
if (qat_uclo_create_batch_init_list(handle, init_mem, ae,
&(obj_handle->lm_init_tab[ae])))
return -EINVAL;
return 0;
}
static int qat_uclo_init_umem_seg(struct icp_qat_fw_loader_handle *handle,
struct icp_qat_uof_initmem *init_mem)
{
struct icp_qat_uclo_objhandle *obj_handle = handle->obj_handle;
unsigned int ae, ustore_size, uaddr, i;
ustore_size = obj_handle->ustore_phy_size;
if (qat_uclo_fetch_initmem_ae(handle, init_mem, ustore_size, &ae))
return -EINVAL;
if (qat_uclo_create_batch_init_list(handle, init_mem, ae,
&(obj_handle->umem_init_tab[ae])))
return -EINVAL;
/* set the highest ustore address referenced */
uaddr = (init_mem->addr + init_mem->num_in_bytes) >> 0x2;
for (i = 0; i < obj_handle->ae_data[ae].slice_num; i++) {
if (obj_handle->ae_data[ae].ae_slices[i].
encap_image->uwords_num < uaddr)
obj_handle->ae_data[ae].ae_slices[i].
encap_image->uwords_num = uaddr;
}
return 0;
}
#define ICP_DH895XCC_PESRAM_BAR_SIZE 0x80000
static int qat_uclo_init_ae_memory(struct icp_qat_fw_loader_handle *handle,
struct icp_qat_uof_initmem *init_mem)
{
unsigned int i;
struct icp_qat_uof_memvar_attr *mem_val_attr;
mem_val_attr =
(struct icp_qat_uof_memvar_attr *)((unsigned long)init_mem +
sizeof(struct icp_qat_uof_initmem));
switch (init_mem->region) {
case ICP_QAT_UOF_SRAM_REGION:
if ((init_mem->addr + init_mem->num_in_bytes) >
ICP_DH895XCC_PESRAM_BAR_SIZE) {
pr_err("QAT: initmem on SRAM is out of range");
return -EINVAL;
}
for (i = 0; i < init_mem->val_attr_num; i++) {
qat_uclo_wr_sram_by_words(handle,
init_mem->addr +
mem_val_attr->offset_in_byte,
&mem_val_attr->value, 4);
mem_val_attr++;
}
break;
case ICP_QAT_UOF_LMEM_REGION:
if (qat_uclo_init_lmem_seg(handle, init_mem))
return -EINVAL;
break;
case ICP_QAT_UOF_UMEM_REGION:
if (qat_uclo_init_umem_seg(handle, init_mem))
return -EINVAL;
break;
default:
pr_err("QAT: initmem region error. region type=0x%x\n",
init_mem->region);
return -EINVAL;
}
return 0;
}
static int qat_uclo_init_ustore(struct icp_qat_fw_loader_handle *handle,
struct icp_qat_uclo_encapme *image)
{
unsigned int i;
struct icp_qat_uclo_encap_page *page;
struct icp_qat_uof_image *uof_image;
unsigned char ae;
unsigned int ustore_size;
unsigned int patt_pos;
struct icp_qat_uclo_objhandle *obj_handle = handle->obj_handle;
uint64_t *fill_data;
uof_image = image->img_ptr;
fill_data = kzalloc(ICP_QAT_UCLO_MAX_USTORE * sizeof(uint64_t),
GFP_KERNEL);
if (!fill_data)
return -EFAULT;
for (i = 0; i < ICP_QAT_UCLO_MAX_USTORE; i++)
memcpy(&fill_data[i], &uof_image->fill_pattern,
sizeof(uint64_t));
page = image->page;
for (ae = 0; ae <= handle->hal_handle->ae_max_num; ae++) {
if (!test_bit(ae, (unsigned long *)&(uof_image->ae_assigned)))
continue;
ustore_size = obj_handle->ae_data[ae].eff_ustore_size;
patt_pos = page->beg_addr_p + page->micro_words_num;
qat_hal_wr_uwords(handle, (unsigned char)ae, 0,
page->beg_addr_p, &fill_data[0]);
qat_hal_wr_uwords(handle, (unsigned char)ae, patt_pos,
ustore_size - patt_pos + 1,
&fill_data[page->beg_addr_p]);
}
kfree(fill_data);
return 0;
}
static int qat_uclo_init_memory(struct icp_qat_fw_loader_handle *handle)
{
unsigned int i;
int status = 0;
struct icp_qat_uclo_objhandle *obj_handle = handle->obj_handle;
struct icp_qat_uof_initmem *initmem = obj_handle->init_mem_tab.init_mem;
int ae;
for (i = 0; i < obj_handle->init_mem_tab.entry_num; i++) {
if (initmem->num_in_bytes) {
if (qat_uclo_init_ae_memory(handle, initmem))
return -EINVAL;
}
initmem = (struct icp_qat_uof_initmem *)((unsigned long)(
(unsigned long)initmem +
sizeof(struct icp_qat_uof_initmem)) +
(sizeof(struct icp_qat_uof_memvar_attr) *
initmem->val_attr_num));
}
for (ae = 0; ae < ICP_QAT_UCLO_MAX_AE; ae++) {
if (qat_hal_batch_wr_lm(handle, ae,
obj_handle->lm_init_tab[ae])) {
pr_err("QAT: fail to batch init lmem for AE %d\n", ae);
return -EINVAL;
}
qat_uclo_cleanup_batch_init_list(handle,
&obj_handle->lm_init_tab[ae]);
qat_uclo_batch_wr_umem(handle, ae,
obj_handle->umem_init_tab[ae]);
qat_uclo_cleanup_batch_init_list(handle,
&obj_handle->
umem_init_tab[ae]);
}
return status;
}
static void *qat_uclo_find_chunk(struct icp_qat_uof_objhdr *obj_hdr,
char *chunk_id, void *cur)
{
int i;
struct icp_qat_uof_chunkhdr *chunk_hdr =
(struct icp_qat_uof_chunkhdr *)
((unsigned long)obj_hdr + sizeof(struct icp_qat_uof_objhdr));
for (i = 0; i < obj_hdr->num_chunks; i++) {
if ((cur < (void *)&chunk_hdr[i]) &&
!(strncmp(chunk_hdr[i].chunk_id, chunk_id,
ICP_QAT_UOF_OBJID_LEN))) {
return &chunk_hdr[i];
}
}
return NULL;
}
static unsigned int qat_uclo_calc_checksum(unsigned int reg, int ch)
{
int i;
unsigned int topbit = 1 << 0xF;
unsigned int inbyte = (unsigned int)((reg >> 0x18) ^ ch);
reg ^= inbyte << 0x8;
for (i = 0; i < 0x8; i++) {
if (reg & topbit)
reg = (reg << 1) ^ 0x1021;
else
reg <<= 1;
}
return reg & 0xFFFF;
}
static unsigned int qat_uclo_calc_str_checksum(char *ptr, int num)
{
unsigned int chksum = 0;
if (ptr)
while (num--)
chksum = qat_uclo_calc_checksum(chksum, *ptr++);
return chksum;
}
static struct icp_qat_uclo_objhdr *
qat_uclo_map_chunk(char *buf, struct icp_qat_uof_filehdr *file_hdr,
char *chunk_id)
{
struct icp_qat_uof_filechunkhdr *file_chunk;
struct icp_qat_uclo_objhdr *obj_hdr;
void *chunk;
int i;
file_chunk = (struct icp_qat_uof_filechunkhdr *)
(buf + sizeof(struct icp_qat_uof_filehdr));
for (i = 0; i < file_hdr->num_chunks; i++) {
if (!(strncmp(file_chunk->chunk_id, chunk_id,
ICP_QAT_UOF_OBJID_LEN))) {
chunk = buf + file_chunk->offset;
if (file_chunk->checksum != qat_uclo_calc_str_checksum(
(char *)chunk, file_chunk->size))
break;
obj_hdr = kzalloc(sizeof(*obj_hdr), GFP_KERNEL);
if (!obj_hdr)
break;
obj_hdr->file_buff = chunk;
obj_hdr->checksum = file_chunk->checksum;
obj_hdr->size = file_chunk->size;
return obj_hdr;
}
file_chunk++;
}
return NULL;
}
static unsigned int
qat_uclo_check_image_compat(struct icp_qat_uof_encap_obj *encap_uof_obj,
struct icp_qat_uof_image *image)
{
struct icp_qat_uof_objtable *uc_var_tab, *imp_var_tab, *imp_expr_tab;
struct icp_qat_uof_objtable *neigh_reg_tab;
struct icp_qat_uof_code_page *code_page;
code_page = (struct icp_qat_uof_code_page *)
((char *)image + sizeof(struct icp_qat_uof_image));
uc_var_tab = (struct icp_qat_uof_objtable *)(encap_uof_obj->beg_uof +
code_page->uc_var_tab_offset);
imp_var_tab = (struct icp_qat_uof_objtable *)(encap_uof_obj->beg_uof +
code_page->imp_var_tab_offset);
imp_expr_tab = (struct icp_qat_uof_objtable *)
(encap_uof_obj->beg_uof +
code_page->imp_expr_tab_offset);
if (uc_var_tab->entry_num || imp_var_tab->entry_num ||
imp_expr_tab->entry_num) {
pr_err("QAT: UOF can't contain imported variable to be parsed");
return -EINVAL;
}
neigh_reg_tab = (struct icp_qat_uof_objtable *)
(encap_uof_obj->beg_uof +
code_page->neigh_reg_tab_offset);
if (neigh_reg_tab->entry_num) {
pr_err("QAT: UOF can't contain shared control store feature");
return -EINVAL;
}
if (image->numpages > 1) {
pr_err("QAT: UOF can't contain multiple pages");
return -EINVAL;
}
if (ICP_QAT_SHARED_USTORE_MODE(image->ae_mode)) {
pr_err("QAT: UOF can't use shared control store feature");
return -EFAULT;
}
if (RELOADABLE_CTX_SHARED_MODE(image->ae_mode)) {
pr_err("QAT: UOF can't use reloadable feature");
return -EFAULT;
}
return 0;
}
static void qat_uclo_map_image_pages(struct icp_qat_uof_encap_obj
*encap_uof_obj,
struct icp_qat_uof_image *img,
struct icp_qat_uclo_encap_page *page)
{
struct icp_qat_uof_code_page *code_page;
struct icp_qat_uof_code_area *code_area;
struct icp_qat_uof_objtable *uword_block_tab;
struct icp_qat_uof_uword_block *uwblock;
int i;
code_page = (struct icp_qat_uof_code_page *)
((char *)img + sizeof(struct icp_qat_uof_image));
page->def_page = code_page->def_page;
page->page_region = code_page->page_region;
page->beg_addr_v = code_page->beg_addr_v;
page->beg_addr_p = code_page->beg_addr_p;
code_area = (struct icp_qat_uof_code_area *)(encap_uof_obj->beg_uof +
code_page->code_area_offset);
page->micro_words_num = code_area->micro_words_num;
uword_block_tab = (struct icp_qat_uof_objtable *)
(encap_uof_obj->beg_uof +
code_area->uword_block_tab);
page->uwblock_num = uword_block_tab->entry_num;
uwblock = (struct icp_qat_uof_uword_block *)((char *)uword_block_tab +
sizeof(struct icp_qat_uof_objtable));
page->uwblock = (struct icp_qat_uclo_encap_uwblock *)uwblock;
for (i = 0; i < uword_block_tab->entry_num; i++)
page->uwblock[i].micro_words =
(unsigned long)encap_uof_obj->beg_uof + uwblock[i].uword_offset;
}
static int qat_uclo_map_uimage(struct icp_qat_uclo_objhandle *obj_handle,
struct icp_qat_uclo_encapme *ae_uimage,
int max_image)
{
int a = 0, i;
struct icp_qat_uof_chunkhdr *chunk_hdr = NULL;
struct icp_qat_uof_image *image;
struct icp_qat_uof_objtable *ae_regtab;
struct icp_qat_uof_objtable *init_reg_sym_tab;
struct icp_qat_uof_objtable *sbreak_tab;
struct icp_qat_uof_encap_obj *encap_uof_obj =
&obj_handle->encap_uof_obj;
for (a = 0; a < max_image; a++) {
chunk_hdr = qat_uclo_find_chunk(encap_uof_obj->obj_hdr,
ICP_QAT_UOF_IMAG, chunk_hdr);
if (!chunk_hdr)
break;
image = (struct icp_qat_uof_image *)(encap_uof_obj->beg_uof +
chunk_hdr->offset);
ae_regtab = (struct icp_qat_uof_objtable *)
(image->reg_tab_offset +
obj_handle->obj_hdr->file_buff);
ae_uimage[a].ae_reg_num = ae_regtab->entry_num;
ae_uimage[a].ae_reg = (struct icp_qat_uof_ae_reg *)
(((char *)ae_regtab) +
sizeof(struct icp_qat_uof_objtable));
init_reg_sym_tab = (struct icp_qat_uof_objtable *)
(image->init_reg_sym_tab +
obj_handle->obj_hdr->file_buff);
ae_uimage[a].init_regsym_num = init_reg_sym_tab->entry_num;
ae_uimage[a].init_regsym = (struct icp_qat_uof_init_regsym *)
(((char *)init_reg_sym_tab) +
sizeof(struct icp_qat_uof_objtable));
sbreak_tab = (struct icp_qat_uof_objtable *)
(image->sbreak_tab + obj_handle->obj_hdr->file_buff);
ae_uimage[a].sbreak_num = sbreak_tab->entry_num;
ae_uimage[a].sbreak = (struct icp_qat_uof_sbreak *)
(((char *)sbreak_tab) +
sizeof(struct icp_qat_uof_objtable));
ae_uimage[a].img_ptr = image;
if (qat_uclo_check_image_compat(encap_uof_obj, image))
goto out_err;
ae_uimage[a].page =
kzalloc(sizeof(struct icp_qat_uclo_encap_page),
GFP_KERNEL);
if (!ae_uimage[a].page)
goto out_err;
qat_uclo_map_image_pages(encap_uof_obj, image,
ae_uimage[a].page);
}
return a;
out_err:
for (i = 0; i < a; i++)
kfree(ae_uimage[i].page);
return 0;
}
static int qat_uclo_map_ae(struct icp_qat_fw_loader_handle *handle, int max_ae)
{
int i, ae;
int mflag = 0;
struct icp_qat_uclo_objhandle *obj_handle = handle->obj_handle;
for (ae = 0; ae <= max_ae; ae++) {
if (!test_bit(ae, (unsigned long *)
&(handle->hal_handle->ae_mask)))
continue;
for (i = 0; i < obj_handle->uimage_num; i++) {
if (!test_bit(ae, (unsigned long *)
&(obj_handle->ae_uimage[i].img_ptr->ae_assigned)))
continue;
mflag = 1;
if (qat_uclo_init_ae_data(obj_handle, ae, i))
return -EINVAL;
}
}
if (!mflag) {
pr_err("QAT: uimage uses AE not set");
return -EINVAL;
}
return 0;
}
static struct icp_qat_uof_strtable *
qat_uclo_map_str_table(struct icp_qat_uclo_objhdr *obj_hdr,
char *tab_name, struct icp_qat_uof_strtable *str_table)
{
struct icp_qat_uof_chunkhdr *chunk_hdr;
chunk_hdr = qat_uclo_find_chunk((struct icp_qat_uof_objhdr *)
obj_hdr->file_buff, tab_name, NULL);
if (chunk_hdr) {
int hdr_size;
memcpy(&str_table->table_len, obj_hdr->file_buff +
chunk_hdr->offset, sizeof(str_table->table_len));
hdr_size = (char *)&str_table->strings - (char *)str_table;
str_table->strings = (unsigned long)obj_hdr->file_buff +
chunk_hdr->offset + hdr_size;
return str_table;
}
return NULL;
}
static void
qat_uclo_map_initmem_table(struct icp_qat_uof_encap_obj *encap_uof_obj,
struct icp_qat_uclo_init_mem_table *init_mem_tab)
{
struct icp_qat_uof_chunkhdr *chunk_hdr;
chunk_hdr = qat_uclo_find_chunk(encap_uof_obj->obj_hdr,
ICP_QAT_UOF_IMEM, NULL);
if (chunk_hdr) {
memmove(&init_mem_tab->entry_num, encap_uof_obj->beg_uof +
chunk_hdr->offset, sizeof(unsigned int));
init_mem_tab->init_mem = (struct icp_qat_uof_initmem *)
(encap_uof_obj->beg_uof + chunk_hdr->offset +
sizeof(unsigned int));
}
}
static int qat_uclo_check_uof_compat(struct icp_qat_uclo_objhandle *obj_handle)
{
unsigned int maj_ver, prod_type = obj_handle->prod_type;
if (!(prod_type & obj_handle->encap_uof_obj.obj_hdr->cpu_type)) {
pr_err("QAT: uof type 0x%x not match with cur platform 0x%x\n",
obj_handle->encap_uof_obj.obj_hdr->cpu_type, prod_type);
return -EINVAL;
}
maj_ver = obj_handle->prod_rev & 0xff;
if ((obj_handle->encap_uof_obj.obj_hdr->max_cpu_ver < maj_ver) ||
(obj_handle->encap_uof_obj.obj_hdr->min_cpu_ver > maj_ver)) {
pr_err("QAT: uof majVer 0x%x out of range\n", maj_ver);
return -EINVAL;
}
return 0;
}
static int qat_uclo_init_reg(struct icp_qat_fw_loader_handle *handle,
unsigned char ae, unsigned char ctx_mask,
enum icp_qat_uof_regtype reg_type,
unsigned short reg_addr, unsigned int value)
{
switch (reg_type) {
case ICP_GPA_ABS:
case ICP_GPB_ABS:
ctx_mask = 0;
case ICP_GPA_REL:
case ICP_GPB_REL:
return qat_hal_init_gpr(handle, ae, ctx_mask, reg_type,
reg_addr, value);
case ICP_SR_ABS:
case ICP_DR_ABS:
case ICP_SR_RD_ABS:
case ICP_DR_RD_ABS:
ctx_mask = 0;
case ICP_SR_REL:
case ICP_DR_REL:
case ICP_SR_RD_REL:
case ICP_DR_RD_REL:
return qat_hal_init_rd_xfer(handle, ae, ctx_mask, reg_type,
reg_addr, value);
case ICP_SR_WR_ABS:
case ICP_DR_WR_ABS:
ctx_mask = 0;
case ICP_SR_WR_REL:
case ICP_DR_WR_REL:
return qat_hal_init_wr_xfer(handle, ae, ctx_mask, reg_type,
reg_addr, value);
case ICP_NEIGH_REL:
return qat_hal_init_nn(handle, ae, ctx_mask, reg_addr, value);
default:
pr_err("QAT: UOF uses not supported reg type 0x%x\n", reg_type);
return -EFAULT;
}
return 0;
}
static int qat_uclo_init_reg_sym(struct icp_qat_fw_loader_handle *handle,
unsigned int ae,
struct icp_qat_uclo_encapme *encap_ae)
{
unsigned int i;
unsigned char ctx_mask;
struct icp_qat_uof_init_regsym *init_regsym;
if (ICP_QAT_CTX_MODE(encap_ae->img_ptr->ae_mode) ==
ICP_QAT_UCLO_MAX_CTX)
ctx_mask = 0xff;
else
ctx_mask = 0x55;
for (i = 0; i < encap_ae->init_regsym_num; i++) {
unsigned int exp_res;
init_regsym = &encap_ae->init_regsym[i];
exp_res = init_regsym->value;
switch (init_regsym->init_type) {
case ICP_QAT_UOF_INIT_REG:
qat_uclo_init_reg(handle, ae, ctx_mask,
(enum icp_qat_uof_regtype)
init_regsym->reg_type,
(unsigned short)init_regsym->reg_addr,
exp_res);
break;
case ICP_QAT_UOF_INIT_REG_CTX:
/* check if ctx is appropriate for the ctxMode */
if (!((1 << init_regsym->ctx) & ctx_mask)) {
pr_err("QAT: invalid ctx num = 0x%x\n",
init_regsym->ctx);
return -EINVAL;
}
qat_uclo_init_reg(handle, ae,
(unsigned char)
(1 << init_regsym->ctx),
(enum icp_qat_uof_regtype)
init_regsym->reg_type,
(unsigned short)init_regsym->reg_addr,
exp_res);
break;
case ICP_QAT_UOF_INIT_EXPR:
pr_err("QAT: INIT_EXPR feature not supported\n");
return -EINVAL;
case ICP_QAT_UOF_INIT_EXPR_ENDIAN_SWAP:
pr_err("QAT: INIT_EXPR_ENDIAN_SWAP feature not supported\n");
return -EINVAL;
default:
break;
}
}
return 0;
}
static int qat_uclo_init_globals(struct icp_qat_fw_loader_handle *handle)
{
struct icp_qat_uclo_objhandle *obj_handle = handle->obj_handle;
unsigned int s, ae;
if (obj_handle->global_inited)
return 0;
if (obj_handle->init_mem_tab.entry_num) {
if (qat_uclo_init_memory(handle)) {
pr_err("QAT: initalize memory failed\n");
return -EINVAL;
}
}
for (ae = 0; ae <= handle->hal_handle->ae_max_num; ae++) {
for (s = 0; s < obj_handle->ae_data[ae].slice_num; s++) {
if (!obj_handle->ae_data[ae].ae_slices[s].encap_image)
continue;
if (qat_uclo_init_reg_sym(handle, ae,
obj_handle->ae_data[ae].
ae_slices[s].encap_image))
return -EINVAL;
}
}
obj_handle->global_inited = 1;
return 0;
}
static int qat_uclo_set_ae_mode(struct icp_qat_fw_loader_handle *handle)
{
unsigned char ae, nn_mode, s;
struct icp_qat_uof_image *uof_image;
struct icp_qat_uclo_aedata *ae_data;
struct icp_qat_uclo_objhandle *obj_handle = handle->obj_handle;
for (ae = 0; ae <= handle->hal_handle->ae_max_num; ae++) {
if (!test_bit(ae,
(unsigned long *)&(handle->hal_handle->ae_mask)))
continue;
ae_data = &(obj_handle->ae_data[ae]);
for (s = 0; s < ae_data->slice_num && s < ICP_QAT_UCLO_MAX_CTX;
s++) {
if (!(obj_handle->ae_data[ae].ae_slices[s].encap_image))
continue;
uof_image = ae_data->ae_slices[s].encap_image->img_ptr;
if (qat_hal_set_ae_ctx_mode(handle, ae,
(char)ICP_QAT_CTX_MODE
(uof_image->ae_mode))) {
pr_err("QAT: qat_hal_set_ae_ctx_mode error\n");
return -EFAULT;
}
nn_mode = ICP_QAT_NN_MODE(uof_image->ae_mode);
if (qat_hal_set_ae_nn_mode(handle, ae, nn_mode)) {
pr_err("QAT: qat_hal_set_ae_nn_mode error\n");
return -EFAULT;
}
if (qat_hal_set_ae_lm_mode(handle, ae, ICP_LMEM0,
(char)ICP_QAT_LOC_MEM0_MODE
(uof_image->ae_mode))) {
pr_err("QAT: qat_hal_set_ae_lm_mode error\n ");
return -EFAULT;
}
if (qat_hal_set_ae_lm_mode(handle, ae, ICP_LMEM1,
(char)ICP_QAT_LOC_MEM1_MODE
(uof_image->ae_mode))) {
pr_err("QAT: qat_hal_set_ae_lm_mode error\n ");
return -EFAULT;
}
}
}
return 0;
}
static void qat_uclo_init_uword_num(struct icp_qat_fw_loader_handle *handle)
{
struct icp_qat_uclo_objhandle *obj_handle = handle->obj_handle;
struct icp_qat_uclo_encapme *image;
int a;
for (a = 0; a < obj_handle->uimage_num; a++) {
image = &obj_handle->ae_uimage[a];
image->uwords_num = image->page->beg_addr_p +
image->page->micro_words_num;
}
}
static int qat_uclo_parse_uof_obj(struct icp_qat_fw_loader_handle *handle)
{
struct icp_qat_uclo_objhandle *obj_handle = handle->obj_handle;
unsigned int ae;
obj_handle->uword_buf = kzalloc(UWORD_CPYBUF_SIZE * sizeof(uint64_t),
GFP_KERNEL);
if (!obj_handle->uword_buf)
return -ENOMEM;
obj_handle->encap_uof_obj.beg_uof = obj_handle->obj_hdr->file_buff;
obj_handle->encap_uof_obj.obj_hdr = (struct icp_qat_uof_objhdr *)
obj_handle->obj_hdr->file_buff;
obj_handle->encap_uof_obj.chunk_hdr = (struct icp_qat_uof_chunkhdr *)
obj_handle->obj_hdr->file_buff + sizeof(struct icp_qat_uof_objhdr);
obj_handle->uword_in_bytes = 6;
obj_handle->prod_type = ICP_QAT_AC_C_CPU_TYPE;
obj_handle->prod_rev = PID_MAJOR_REV |
(PID_MINOR_REV & handle->hal_handle->revision_id);
if (qat_uclo_check_uof_compat(obj_handle)) {
pr_err("QAT: uof incompatible\n ");
return -EINVAL;
}
obj_handle->ustore_phy_size = ICP_QAT_UCLO_MAX_USTORE;
if (!(obj_handle->obj_hdr->file_buff) ||
!(qat_uclo_map_str_table(obj_handle->obj_hdr, ICP_QAT_UOF_STRT,
&(obj_handle->str_table)))) {
pr_err("QAT: uof doesn't have effective images");
goto out_err;
}
obj_handle->uimage_num =
qat_uclo_map_uimage(obj_handle, obj_handle->ae_uimage,
ICP_QAT_UCLO_MAX_AE * ICP_QAT_UCLO_MAX_CTX);
if (!obj_handle->uimage_num)
goto out_err;
if (qat_uclo_map_ae(handle, handle->hal_handle->ae_max_num)) {
pr_err("QAT: Bad object\n ");
goto out_check_uof_aemask_err;
}
qat_uclo_init_uword_num(handle);
qat_uclo_map_initmem_table(&(obj_handle->encap_uof_obj),
&(obj_handle->init_mem_tab));
if (qat_uclo_set_ae_mode(handle))
goto out_check_uof_aemask_err;
return 0;
out_check_uof_aemask_err:
for (ae = 0; ae < obj_handle->uimage_num; ae++)
kfree(obj_handle->ae_uimage[ae].page);
out_err:
kfree(obj_handle->uword_buf);
return -EFAULT;
}
int qat_uclo_map_uof_obj(struct icp_qat_fw_loader_handle *handle,
void *addr_ptr, int mem_size)
{
struct icp_qat_uof_filehdr *filehdr;
struct icp_qat_uclo_objhandle *objhdl;
BUILD_BUG_ON(ICP_QAT_UCLO_MAX_AE >=
(sizeof(handle->hal_handle->ae_mask) * 8));
if (!handle || !addr_ptr || mem_size < 24)
return -EINVAL;
objhdl = kzalloc(sizeof(*objhdl), GFP_KERNEL);
if (!objhdl)
return -ENOMEM;
objhdl->obj_buf = kmemdup(addr_ptr, mem_size, GFP_KERNEL);
if (!objhdl->obj_buf)
goto out_objbuf_err;
filehdr = (struct icp_qat_uof_filehdr *)objhdl->obj_buf;
if (qat_uclo_check_format(filehdr))
goto out_objhdr_err;
objhdl->obj_hdr = qat_uclo_map_chunk((char *)objhdl->obj_buf, filehdr,
ICP_QAT_UOF_OBJS);
if (!objhdl->obj_hdr) {
pr_err("QAT: object file chunk is null\n");
goto out_objhdr_err;
}
handle->obj_handle = objhdl;
if (qat_uclo_parse_uof_obj(handle))
goto out_overlay_obj_err;
return 0;
out_overlay_obj_err:
handle->obj_handle = NULL;
kfree(objhdl->obj_hdr);
out_objhdr_err:
kfree(objhdl->obj_buf);
out_objbuf_err:
kfree(objhdl);
return -ENOMEM;
}
int qat_uclo_del_uof_obj(struct icp_qat_fw_loader_handle *handle)
{
struct icp_qat_uclo_objhandle *obj_handle = handle->obj_handle;
int a;
kfree(obj_handle->uword_buf);
for (a = 0; a < obj_handle->uimage_num; a++)
kfree(obj_handle->ae_uimage[a].page);
for (a = 0; a <= (int)handle->hal_handle->ae_max_num; a++)
qat_uclo_free_ae_data(&obj_handle->ae_data[a]);
kfree(obj_handle->obj_hdr);
kfree(obj_handle->obj_buf);
kfree(obj_handle);
handle->obj_handle = NULL;
return 0;
}
static void qat_uclo_fill_uwords(struct icp_qat_uclo_objhandle *obj_handle,
struct icp_qat_uclo_encap_page *encap_page,
uint64_t *uword, unsigned int addr_p,
unsigned int raddr, uint64_t fill)
{
uint64_t uwrd = 0;
unsigned int i;
if (!encap_page) {
*uword = fill;
return;
}
for (i = 0; i < encap_page->uwblock_num; i++) {
if (raddr >= encap_page->uwblock[i].start_addr &&
raddr <= encap_page->uwblock[i].start_addr +
encap_page->uwblock[i].words_num - 1) {
raddr -= encap_page->uwblock[i].start_addr;
raddr *= obj_handle->uword_in_bytes;
memcpy(&uwrd, (void *)(((unsigned long)
encap_page->uwblock[i].micro_words) + raddr),
obj_handle->uword_in_bytes);
uwrd = uwrd & 0xbffffffffffull;
}
}
*uword = uwrd;
if (*uword == INVLD_UWORD)
*uword = fill;
}
static int qat_uclo_wr_uimage_raw_page(struct icp_qat_fw_loader_handle *handle,
struct icp_qat_uclo_encap_page
*encap_page, unsigned int ae)
{
unsigned int uw_physical_addr, uw_relative_addr, i, words_num, cpylen;
struct icp_qat_uclo_objhandle *obj_handle = handle->obj_handle;
uint64_t fill_pat;
int status = 0;
/* load the page starting at appropriate ustore address */
/* get fill-pattern from an image -- they are all the same */
memcpy(&fill_pat, obj_handle->ae_uimage[0].img_ptr->fill_pattern,
sizeof(uint64_t));
uw_physical_addr = encap_page->beg_addr_p;
uw_relative_addr = 0;
words_num = encap_page->micro_words_num;
while (words_num) {
if (words_num < UWORD_CPYBUF_SIZE)
cpylen = words_num;
else
cpylen = UWORD_CPYBUF_SIZE;
/* load the buffer */
for (i = 0; i < cpylen; i++)
qat_uclo_fill_uwords(obj_handle, encap_page,
&obj_handle->uword_buf[i],
uw_physical_addr + i,
uw_relative_addr + i, fill_pat);
/* copy the buffer to ustore */
qat_hal_wr_uwords(handle, (unsigned char)ae,
uw_physical_addr, cpylen,
obj_handle->uword_buf);
uw_physical_addr += cpylen;
uw_relative_addr += cpylen;
words_num -= cpylen;
}
return status;
}
static int
qat_uclo_wr_uimage_pages(struct icp_qat_fw_loader_handle *handle,
struct icp_qat_uof_image *image)
{
struct icp_qat_uclo_objhandle *obj_handle = handle->obj_handle;
unsigned int ctx_mask, s;
struct icp_qat_uclo_page *page;
unsigned char ae;
int retval = 0;
int ctx;
if (ICP_QAT_CTX_MODE(image->ae_mode) == ICP_QAT_UCLO_MAX_CTX)
ctx_mask = 0xff;
else
ctx_mask = 0x55;
/* load the default page and set assigned CTX PC
* to the entrypoint address */
for (ae = 0; ae <= handle->hal_handle->ae_max_num; ae++) {
if (!test_bit(ae, (unsigned long *)&(image->ae_assigned)))
continue;
/* find the slice to which this image is assigned */
for (s = 0; s < obj_handle->ae_data[ae].slice_num; s++) {
if (image->ctx_assigned & obj_handle->ae_data[ae].
ae_slices[s].ctx_mask_assigned)
break;
}
if (s >= obj_handle->ae_data[ae].slice_num)
continue;
page = obj_handle->ae_data[ae].ae_slices[s].page;
if (!page->encap_page->def_page)
continue;
if (qat_uclo_wr_uimage_raw_page(handle, page->encap_page, ae))
return -EINVAL;
page = obj_handle->ae_data[ae].ae_slices[s].page;
for (ctx = 0; ctx < ICP_QAT_UCLO_MAX_CTX; ctx++)
obj_handle->ae_data[ae].ae_slices[s].cur_page[ctx] =
(ctx_mask & (1 << ctx)) ? page : NULL;
qat_hal_set_live_ctx(handle, (unsigned char)ae,
image->ctx_assigned);
qat_hal_set_pc(handle, (unsigned char)ae, image->ctx_assigned,
image->entry_address);
}
return retval;
}
int qat_uclo_wr_all_uimage(struct icp_qat_fw_loader_handle *handle)
{
struct icp_qat_uclo_objhandle *obj_handle = handle->obj_handle;
unsigned int i;
if (qat_uclo_init_globals(handle))
return -EINVAL;
for (i = 0; i < obj_handle->uimage_num; i++) {
if (!(obj_handle->ae_uimage[i].img_ptr))
return -EINVAL;
if (qat_uclo_init_ustore(handle, &(obj_handle->ae_uimage[i])))
return -EINVAL;
if (qat_uclo_wr_uimage_pages(handle,
obj_handle->ae_uimage[i].img_ptr))
return -EINVAL;
}
return 0;
}
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment