Commit d98793b5 authored by Dave Jiang's avatar Dave Jiang Committed by Vinod Koul

dmaengine: idxd: fix wq config registers offset programming

DSA spec v1.1 [1] updated to include a stride size register for WQ
configuration that will specify how much space is reserved for the WQ
configuration register set. This change is expected to be in the final
gen1 DSA hardware. Fix the driver to use WQCFG_OFFSET() for all WQ
offset calculation and fixup WQCFG_OFFSET() to use the new calculated
wq size.

[1]: https://software.intel.com/content/www/us/en/develop/download/intel-data-streaming-accelerator-preliminary-architecture-specification.html

Fixes: bfe1d560 ("dmaengine: idxd: Init and probe for Intel data accelerators")
Signed-off-by: default avatarDave Jiang <dave.jiang@intel.com>
Link: https://lore.kernel.org/r/160383444959.48058.14249265538404901781.stgit@djiang5-desk3.ch.intel.comSigned-off-by: default avatarVinod Koul <vkoul@kernel.org>
parent 3650b228
...@@ -295,7 +295,7 @@ void idxd_wq_disable_cleanup(struct idxd_wq *wq) ...@@ -295,7 +295,7 @@ void idxd_wq_disable_cleanup(struct idxd_wq *wq)
int i, wq_offset; int i, wq_offset;
lockdep_assert_held(&idxd->dev_lock); lockdep_assert_held(&idxd->dev_lock);
memset(&wq->wqcfg, 0, sizeof(wq->wqcfg)); memset(wq->wqcfg, 0, idxd->wqcfg_size);
wq->type = IDXD_WQT_NONE; wq->type = IDXD_WQT_NONE;
wq->size = 0; wq->size = 0;
wq->group = NULL; wq->group = NULL;
...@@ -304,8 +304,8 @@ void idxd_wq_disable_cleanup(struct idxd_wq *wq) ...@@ -304,8 +304,8 @@ void idxd_wq_disable_cleanup(struct idxd_wq *wq)
clear_bit(WQ_FLAG_DEDICATED, &wq->flags); clear_bit(WQ_FLAG_DEDICATED, &wq->flags);
memset(wq->name, 0, WQ_NAME_SIZE); memset(wq->name, 0, WQ_NAME_SIZE);
for (i = 0; i < 8; i++) { for (i = 0; i < WQCFG_STRIDES(idxd); i++) {
wq_offset = idxd->wqcfg_offset + wq->id * 32 + i * sizeof(u32); wq_offset = WQCFG_OFFSET(idxd, wq->id, i);
iowrite32(0, idxd->reg_base + wq_offset); iowrite32(0, idxd->reg_base + wq_offset);
dev_dbg(dev, "WQ[%d][%d][%#x]: %#x\n", dev_dbg(dev, "WQ[%d][%d][%#x]: %#x\n",
wq->id, i, wq_offset, wq->id, i, wq_offset,
...@@ -539,10 +539,10 @@ static int idxd_wq_config_write(struct idxd_wq *wq) ...@@ -539,10 +539,10 @@ static int idxd_wq_config_write(struct idxd_wq *wq)
if (!wq->group) if (!wq->group)
return 0; return 0;
memset(&wq->wqcfg, 0, sizeof(union wqcfg)); memset(wq->wqcfg, 0, idxd->wqcfg_size);
/* byte 0-3 */ /* byte 0-3 */
wq->wqcfg.wq_size = wq->size; wq->wqcfg->wq_size = wq->size;
if (wq->size == 0) { if (wq->size == 0) {
dev_warn(dev, "Incorrect work queue size: 0\n"); dev_warn(dev, "Incorrect work queue size: 0\n");
...@@ -550,22 +550,21 @@ static int idxd_wq_config_write(struct idxd_wq *wq) ...@@ -550,22 +550,21 @@ static int idxd_wq_config_write(struct idxd_wq *wq)
} }
/* bytes 4-7 */ /* bytes 4-7 */
wq->wqcfg.wq_thresh = wq->threshold; wq->wqcfg->wq_thresh = wq->threshold;
/* byte 8-11 */ /* byte 8-11 */
wq->wqcfg.priv = !!(wq->type == IDXD_WQT_KERNEL); wq->wqcfg->priv = !!(wq->type == IDXD_WQT_KERNEL);
wq->wqcfg.mode = 1; wq->wqcfg->mode = 1;
wq->wqcfg->priority = wq->priority;
wq->wqcfg.priority = wq->priority;
/* bytes 12-15 */ /* bytes 12-15 */
wq->wqcfg.max_xfer_shift = ilog2(wq->max_xfer_bytes); wq->wqcfg->max_xfer_shift = ilog2(wq->max_xfer_bytes);
wq->wqcfg.max_batch_shift = ilog2(wq->max_batch_size); wq->wqcfg->max_batch_shift = ilog2(wq->max_batch_size);
dev_dbg(dev, "WQ %d CFGs\n", wq->id); dev_dbg(dev, "WQ %d CFGs\n", wq->id);
for (i = 0; i < 8; i++) { for (i = 0; i < WQCFG_STRIDES(idxd); i++) {
wq_offset = idxd->wqcfg_offset + wq->id * 32 + i * sizeof(u32); wq_offset = WQCFG_OFFSET(idxd, wq->id, i);
iowrite32(wq->wqcfg.bits[i], idxd->reg_base + wq_offset); iowrite32(wq->wqcfg->bits[i], idxd->reg_base + wq_offset);
dev_dbg(dev, "WQ[%d][%d][%#x]: %#x\n", dev_dbg(dev, "WQ[%d][%d][%#x]: %#x\n",
wq->id, i, wq_offset, wq->id, i, wq_offset,
ioread32(idxd->reg_base + wq_offset)); ioread32(idxd->reg_base + wq_offset));
......
...@@ -103,7 +103,7 @@ struct idxd_wq { ...@@ -103,7 +103,7 @@ struct idxd_wq {
u32 priority; u32 priority;
enum idxd_wq_state state; enum idxd_wq_state state;
unsigned long flags; unsigned long flags;
union wqcfg wqcfg; union wqcfg *wqcfg;
u32 vec_ptr; /* interrupt steering */ u32 vec_ptr; /* interrupt steering */
struct dsa_hw_desc **hw_descs; struct dsa_hw_desc **hw_descs;
int num_descs; int num_descs;
...@@ -183,6 +183,7 @@ struct idxd_device { ...@@ -183,6 +183,7 @@ struct idxd_device {
int max_wq_size; int max_wq_size;
int token_limit; int token_limit;
int nr_tokens; /* non-reserved tokens */ int nr_tokens; /* non-reserved tokens */
unsigned int wqcfg_size;
union sw_err_reg sw_err; union sw_err_reg sw_err;
wait_queue_head_t cmd_waitq; wait_queue_head_t cmd_waitq;
......
...@@ -178,6 +178,9 @@ static int idxd_setup_internals(struct idxd_device *idxd) ...@@ -178,6 +178,9 @@ static int idxd_setup_internals(struct idxd_device *idxd)
wq->idxd_cdev.minor = -1; wq->idxd_cdev.minor = -1;
wq->max_xfer_bytes = idxd->max_xfer_bytes; wq->max_xfer_bytes = idxd->max_xfer_bytes;
wq->max_batch_size = idxd->max_batch_size; wq->max_batch_size = idxd->max_batch_size;
wq->wqcfg = devm_kzalloc(dev, idxd->wqcfg_size, GFP_KERNEL);
if (!wq->wqcfg)
return -ENOMEM;
} }
for (i = 0; i < idxd->max_engines; i++) { for (i = 0; i < idxd->max_engines; i++) {
...@@ -251,6 +254,8 @@ static void idxd_read_caps(struct idxd_device *idxd) ...@@ -251,6 +254,8 @@ static void idxd_read_caps(struct idxd_device *idxd)
dev_dbg(dev, "total workqueue size: %u\n", idxd->max_wq_size); dev_dbg(dev, "total workqueue size: %u\n", idxd->max_wq_size);
idxd->max_wqs = idxd->hw.wq_cap.num_wqs; idxd->max_wqs = idxd->hw.wq_cap.num_wqs;
dev_dbg(dev, "max workqueues: %u\n", idxd->max_wqs); dev_dbg(dev, "max workqueues: %u\n", idxd->max_wqs);
idxd->wqcfg_size = 1 << (idxd->hw.wq_cap.wqcfg_size + IDXD_WQCFG_MIN);
dev_dbg(dev, "wqcfg size: %u\n", idxd->wqcfg_size);
/* reading operation capabilities */ /* reading operation capabilities */
for (i = 0; i < 4; i++) { for (i = 0; i < 4; i++) {
......
...@@ -43,7 +43,8 @@ union wq_cap_reg { ...@@ -43,7 +43,8 @@ union wq_cap_reg {
struct { struct {
u64 total_wq_size:16; u64 total_wq_size:16;
u64 num_wqs:8; u64 num_wqs:8;
u64 rsvd:24; u64 wqcfg_size:4;
u64 rsvd:20;
u64 shared_mode:1; u64 shared_mode:1;
u64 dedicated_mode:1; u64 dedicated_mode:1;
u64 rsvd2:1; u64 rsvd2:1;
...@@ -55,6 +56,7 @@ union wq_cap_reg { ...@@ -55,6 +56,7 @@ union wq_cap_reg {
u64 bits; u64 bits;
} __packed; } __packed;
#define IDXD_WQCAP_OFFSET 0x20 #define IDXD_WQCAP_OFFSET 0x20
#define IDXD_WQCFG_MIN 5
union group_cap_reg { union group_cap_reg {
struct { struct {
...@@ -333,4 +335,23 @@ union wqcfg { ...@@ -333,4 +335,23 @@ union wqcfg {
}; };
u32 bits[8]; u32 bits[8];
} __packed; } __packed;
/*
* This macro calculates the offset into the WQCFG register
* idxd - struct idxd *
* n - wq id
* ofs - the index of the 32b dword for the config register
*
* The WQCFG register block is divided into groups per each wq. The n index
* allows us to move to the register group that's for that particular wq.
* Each register is 32bits. The ofs gives us the number of register to access.
*/
#define WQCFG_OFFSET(_idxd_dev, n, ofs) \
({\
typeof(_idxd_dev) __idxd_dev = (_idxd_dev); \
(__idxd_dev)->wqcfg_offset + (n) * (__idxd_dev)->wqcfg_size + sizeof(u32) * (ofs); \
})
#define WQCFG_STRIDES(_idxd_dev) ((_idxd_dev)->wqcfg_size / sizeof(u32))
#endif #endif
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment