2023-10-11 12:36:12

by Damian Muszynski

[permalink] [raw]
Subject: [PATCH 00/11] crypto: qat - add rate limiting feature to qat_4xxx

This set introduces the rate limiting feature to the Intel QAT accelerator.
The Rate Limiting allows control of the rate of the requests that can be
submitted on a ring pair (RP). This allows sharing a QAT device among
multiple users while ensuring a guaranteed throughput.

The first six commits address minor reworks and additions necessary
for the feature's implementation.
The subsequent patch introduces a mechanism for retrieving firmware
feature capabilities.
Patch nr 8 implements the core of the rate limiting.
The final three commits add the required sysfs interface
for configuring this newly introduced feature.

Ciunas Bennett (3):
crypto: qat - add rate limiting sysfs interface
crypto: qat - add rp2svc sysfs attribute
crypto: qat - add num_rps sysfs attribute

Damian Muszynski (4):
units: Add BYTES_PER_*BIT
crypto: qat - add bits.h to icp_qat_hw.h
crypto: qat - add retrieval of fw capabilities
crypto: qat - add rate limiting feature to qat_4xxx

Giovanni Cabiddu (4):
crypto: qat - refactor fw config related functions
crypto: qat - use masks for AE groups
crypto: qat - fix ring to service map for QAT GEN4
crypto: qat - move admin api

Documentation/ABI/testing/sysfs-driver-qat | 46 +
Documentation/ABI/testing/sysfs-driver-qat_rl | 227 ++++
.../intel/qat/qat_4xxx/adf_4xxx_hw_data.c | 190 ++-
.../intel/qat/qat_4xxx/adf_4xxx_hw_data.h | 13 +-
.../intel/qat/qat_c3xxx/adf_c3xxx_hw_data.c | 1 +
.../intel/qat/qat_c62x/adf_c62x_hw_data.c | 1 +
drivers/crypto/intel/qat/qat_common/Makefile | 3 +
.../intel/qat/qat_common/adf_accel_devices.h | 11 +
.../crypto/intel/qat/qat_common/adf_admin.c | 73 +
.../crypto/intel/qat/qat_common/adf_admin.h | 27 +
.../crypto/intel/qat/qat_common/adf_clock.c | 1 +
.../intel/qat/qat_common/adf_cnv_dbgfs.c | 1 +
.../intel/qat/qat_common/adf_common_drv.h | 10 -
.../intel/qat/qat_common/adf_fw_counters.c | 1 +
.../intel/qat/qat_common/adf_gen4_hw_data.h | 7 +
.../crypto/intel/qat/qat_common/adf_gen4_pm.c | 1 +
.../qat/qat_common/adf_gen4_pm_debugfs.c | 1 +
.../intel/qat/qat_common/adf_gen4_timer.c | 1 +
.../intel/qat/qat_common/adf_heartbeat.c | 1 +
.../qat/qat_common/adf_heartbeat_dbgfs.c | 1 +
.../crypto/intel/qat/qat_common/adf_init.c | 12 +
drivers/crypto/intel/qat/qat_common/adf_rl.c | 1186 +++++++++++++++++
drivers/crypto/intel/qat/qat_common/adf_rl.h | 176 +++
.../intel/qat/qat_common/adf_rl_admin.c | 98 ++
.../intel/qat/qat_common/adf_rl_admin.h | 18 +
.../crypto/intel/qat/qat_common/adf_sysfs.c | 80 ++
.../intel/qat/qat_common/adf_sysfs_rl.c | 451 +++++++
.../intel/qat/qat_common/adf_sysfs_rl.h | 11 +
.../qat/qat_common/icp_qat_fw_init_admin.h | 41 +
.../crypto/intel/qat/qat_common/icp_qat_hw.h | 2 +
.../qat/qat_dh895xcc/adf_dh895xcc_hw_data.c | 1 +
include/linux/units.h | 4 +
32 files changed, 2624 insertions(+), 73 deletions(-)
create mode 100644 Documentation/ABI/testing/sysfs-driver-qat_rl
create mode 100644 drivers/crypto/intel/qat/qat_common/adf_admin.h
create mode 100644 drivers/crypto/intel/qat/qat_common/adf_rl.c
create mode 100644 drivers/crypto/intel/qat/qat_common/adf_rl.h
create mode 100644 drivers/crypto/intel/qat/qat_common/adf_rl_admin.c
create mode 100644 drivers/crypto/intel/qat/qat_common/adf_rl_admin.h
create mode 100644 drivers/crypto/intel/qat/qat_common/adf_sysfs_rl.c
create mode 100644 drivers/crypto/intel/qat/qat_common/adf_sysfs_rl.h

--
2.41.0


2023-10-11 12:36:20

by Damian Muszynski

[permalink] [raw]
Subject: [PATCH 01/11] crypto: qat - refactor fw config related functions

From: Giovanni Cabiddu <[email protected]>

The logic that selects the correct adf_fw_config structure based on the
configured service is replicated twice in the uof_get_name() and
uof_get_ae_mask() functions. Refactor the code so that there is no
replication.

This does not introduce any functional change.

Signed-off-by: Giovanni Cabiddu <[email protected]>
Reviewed-by: Damian Muszynski <[email protected]>
---
.../intel/qat/qat_4xxx/adf_4xxx_hw_data.c | 69 ++++++++-----------
1 file changed, 28 insertions(+), 41 deletions(-)

diff --git a/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c b/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
index 12b5d1819111..10839269c4d3 100644
--- a/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
+++ b/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
@@ -394,40 +394,42 @@ static u32 uof_get_num_objs(void)
return ARRAY_SIZE(adf_fw_cy_config);
}

-static const char *uof_get_name(struct adf_accel_dev *accel_dev, u32 obj_num,
- const char * const fw_objs[], int num_objs)
+static const struct adf_fw_config *get_fw_config(struct adf_accel_dev *accel_dev)
{
- int id;
-
switch (get_service_enabled(accel_dev)) {
case SVC_CY:
case SVC_CY2:
- id = adf_fw_cy_config[obj_num].obj;
- break;
+ return adf_fw_cy_config;
case SVC_DC:
- id = adf_fw_dc_config[obj_num].obj;
- break;
+ return adf_fw_dc_config;
case SVC_DCC:
- id = adf_fw_dcc_config[obj_num].obj;
- break;
+ return adf_fw_dcc_config;
case SVC_SYM:
- id = adf_fw_sym_config[obj_num].obj;
- break;
+ return adf_fw_sym_config;
case SVC_ASYM:
- id = adf_fw_asym_config[obj_num].obj;
- break;
+ return adf_fw_asym_config;
case SVC_ASYM_DC:
case SVC_DC_ASYM:
- id = adf_fw_asym_dc_config[obj_num].obj;
- break;
+ return adf_fw_asym_dc_config;
case SVC_SYM_DC:
case SVC_DC_SYM:
- id = adf_fw_sym_dc_config[obj_num].obj;
- break;
+ return adf_fw_sym_dc_config;
default:
- id = -EINVAL;
- break;
+ return NULL;
}
+}
+
+static const char *uof_get_name(struct adf_accel_dev *accel_dev, u32 obj_num,
+ const char * const fw_objs[], int num_objs)
+{
+ const struct adf_fw_config *fw_config;
+ int id;
+
+ fw_config = get_fw_config(accel_dev);
+ if (fw_config)
+ id = fw_config[obj_num].obj;
+ else
+ id = -EINVAL;

if (id < 0 || id > num_objs)
return NULL;
@@ -451,28 +453,13 @@ static const char *uof_get_name_402xx(struct adf_accel_dev *accel_dev, u32 obj_n

static u32 uof_get_ae_mask(struct adf_accel_dev *accel_dev, u32 obj_num)
{
- switch (get_service_enabled(accel_dev)) {
- case SVC_CY:
- return adf_fw_cy_config[obj_num].ae_mask;
- case SVC_DC:
- return adf_fw_dc_config[obj_num].ae_mask;
- case SVC_DCC:
- return adf_fw_dcc_config[obj_num].ae_mask;
- case SVC_CY2:
- return adf_fw_cy_config[obj_num].ae_mask;
- case SVC_SYM:
- return adf_fw_sym_config[obj_num].ae_mask;
- case SVC_ASYM:
- return adf_fw_asym_config[obj_num].ae_mask;
- case SVC_ASYM_DC:
- case SVC_DC_ASYM:
- return adf_fw_asym_dc_config[obj_num].ae_mask;
- case SVC_SYM_DC:
- case SVC_DC_SYM:
- return adf_fw_sym_dc_config[obj_num].ae_mask;
- default:
+ const struct adf_fw_config *fw_config;
+
+ fw_config = get_fw_config(accel_dev);
+ if (!fw_config)
return 0;
- }
+
+ return fw_config[obj_num].ae_mask;
}

void adf_init_hw_data_4xxx(struct adf_hw_device_data *hw_data, u32 dev_id)
--
2.41.0

2023-10-11 12:36:24

by Damian Muszynski

[permalink] [raw]
Subject: [PATCH 02/11] crypto: qat - use masks for AE groups

From: Giovanni Cabiddu <[email protected]>

The adf_fw_config structures hardcode a bit mask that represents the
acceleration engines (AEs) where a certain firmware image will have to
be loaded to. Remove the hardcoded masks and replace them with defines.

This does not introduce any functional change.

Signed-off-by: Giovanni Cabiddu <[email protected]>
Reviewed-by: Damian Muszynski <[email protected]>
---
.../intel/qat/qat_4xxx/adf_4xxx_hw_data.c | 46 ++++++++++---------
1 file changed, 25 insertions(+), 21 deletions(-)

diff --git a/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c b/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
index 10839269c4d3..44b732fb80bc 100644
--- a/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
+++ b/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
@@ -14,6 +14,10 @@
#include "adf_cfg_services.h"
#include "icp_qat_hw.h"

+#define ADF_AE_GROUP_0 GENMASK(3, 0)
+#define ADF_AE_GROUP_1 GENMASK(7, 4)
+#define ADF_AE_GROUP_2 BIT(8)
+
enum adf_fw_objs {
ADF_FW_SYM_OBJ,
ADF_FW_ASYM_OBJ,
@@ -41,45 +45,45 @@ struct adf_fw_config {
};

static const struct adf_fw_config adf_fw_cy_config[] = {
- {0xF0, ADF_FW_SYM_OBJ},
- {0xF, ADF_FW_ASYM_OBJ},
- {0x100, ADF_FW_ADMIN_OBJ},
+ {ADF_AE_GROUP_1, ADF_FW_SYM_OBJ},
+ {ADF_AE_GROUP_0, ADF_FW_ASYM_OBJ},
+ {ADF_AE_GROUP_2, ADF_FW_ADMIN_OBJ},
};

static const struct adf_fw_config adf_fw_dc_config[] = {
- {0xF0, ADF_FW_DC_OBJ},
- {0xF, ADF_FW_DC_OBJ},
- {0x100, ADF_FW_ADMIN_OBJ},
+ {ADF_AE_GROUP_1, ADF_FW_DC_OBJ},
+ {ADF_AE_GROUP_0, ADF_FW_DC_OBJ},
+ {ADF_AE_GROUP_2, ADF_FW_ADMIN_OBJ},
};

static const struct adf_fw_config adf_fw_sym_config[] = {
- {0xF0, ADF_FW_SYM_OBJ},
- {0xF, ADF_FW_SYM_OBJ},
- {0x100, ADF_FW_ADMIN_OBJ},
+ {ADF_AE_GROUP_1, ADF_FW_SYM_OBJ},
+ {ADF_AE_GROUP_0, ADF_FW_SYM_OBJ},
+ {ADF_AE_GROUP_2, ADF_FW_ADMIN_OBJ},
};

static const struct adf_fw_config adf_fw_asym_config[] = {
- {0xF0, ADF_FW_ASYM_OBJ},
- {0xF, ADF_FW_ASYM_OBJ},
- {0x100, ADF_FW_ADMIN_OBJ},
+ {ADF_AE_GROUP_1, ADF_FW_ASYM_OBJ},
+ {ADF_AE_GROUP_0, ADF_FW_ASYM_OBJ},
+ {ADF_AE_GROUP_2, ADF_FW_ADMIN_OBJ},
};

static const struct adf_fw_config adf_fw_asym_dc_config[] = {
- {0xF0, ADF_FW_ASYM_OBJ},
- {0xF, ADF_FW_DC_OBJ},
- {0x100, ADF_FW_ADMIN_OBJ},
+ {ADF_AE_GROUP_1, ADF_FW_ASYM_OBJ},
+ {ADF_AE_GROUP_0, ADF_FW_DC_OBJ},
+ {ADF_AE_GROUP_2, ADF_FW_ADMIN_OBJ},
};

static const struct adf_fw_config adf_fw_sym_dc_config[] = {
- {0xF0, ADF_FW_SYM_OBJ},
- {0xF, ADF_FW_DC_OBJ},
- {0x100, ADF_FW_ADMIN_OBJ},
+ {ADF_AE_GROUP_1, ADF_FW_SYM_OBJ},
+ {ADF_AE_GROUP_0, ADF_FW_DC_OBJ},
+ {ADF_AE_GROUP_2, ADF_FW_ADMIN_OBJ},
};

static const struct adf_fw_config adf_fw_dcc_config[] = {
- {0xF0, ADF_FW_DC_OBJ},
- {0xF, ADF_FW_SYM_OBJ},
- {0x100, ADF_FW_ADMIN_OBJ},
+ {ADF_AE_GROUP_1, ADF_FW_DC_OBJ},
+ {ADF_AE_GROUP_0, ADF_FW_SYM_OBJ},
+ {ADF_AE_GROUP_2, ADF_FW_ADMIN_OBJ},
};

static_assert(ARRAY_SIZE(adf_fw_cy_config) == ARRAY_SIZE(adf_fw_dc_config));
--
2.41.0

2023-10-11 12:36:28

by Damian Muszynski

[permalink] [raw]
Subject: [PATCH 04/11] crypto: qat - move admin api

From: Giovanni Cabiddu <[email protected]>

The admin API is growing and deserves its own include.
Move it from adf_common_drv.h to adf_admin.h.

Signed-off-by: Giovanni Cabiddu <[email protected]>
Reviewed-by: Damian Muszynski <[email protected]>
---
.../intel/qat/qat_4xxx/adf_4xxx_hw_data.c | 1 +
.../intel/qat/qat_c3xxx/adf_c3xxx_hw_data.c | 1 +
.../intel/qat/qat_c62x/adf_c62x_hw_data.c | 1 +
.../crypto/intel/qat/qat_common/adf_admin.c | 1 +
.../crypto/intel/qat/qat_common/adf_admin.h | 19 +++++++++++++++++++
.../crypto/intel/qat/qat_common/adf_clock.c | 1 +
.../intel/qat/qat_common/adf_cnv_dbgfs.c | 1 +
.../intel/qat/qat_common/adf_common_drv.h | 10 ----------
.../intel/qat/qat_common/adf_fw_counters.c | 1 +
.../crypto/intel/qat/qat_common/adf_gen4_pm.c | 1 +
.../qat/qat_common/adf_gen4_pm_debugfs.c | 1 +
.../intel/qat/qat_common/adf_gen4_timer.c | 1 +
.../intel/qat/qat_common/adf_heartbeat.c | 1 +
.../qat/qat_common/adf_heartbeat_dbgfs.c | 1 +
.../qat/qat_dh895xcc/adf_dh895xcc_hw_data.c | 1 +
15 files changed, 32 insertions(+), 10 deletions(-)
create mode 100644 drivers/crypto/intel/qat/qat_common/adf_admin.h

diff --git a/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c b/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
index a5691ba0b724..8a80701b7791 100644
--- a/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
+++ b/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
@@ -2,6 +2,7 @@
/* Copyright(c) 2020 - 2021 Intel Corporation */
#include <linux/iopoll.h>
#include <adf_accel_devices.h>
+#include <adf_admin.h>
#include <adf_cfg.h>
#include <adf_clock.h>
#include <adf_common_drv.h>
diff --git a/drivers/crypto/intel/qat/qat_c3xxx/adf_c3xxx_hw_data.c b/drivers/crypto/intel/qat/qat_c3xxx/adf_c3xxx_hw_data.c
index 9c00c441b602..a882e0ea2279 100644
--- a/drivers/crypto/intel/qat/qat_c3xxx/adf_c3xxx_hw_data.c
+++ b/drivers/crypto/intel/qat/qat_c3xxx/adf_c3xxx_hw_data.c
@@ -1,6 +1,7 @@
// SPDX-License-Identifier: (BSD-3-Clause OR GPL-2.0-only)
/* Copyright(c) 2014 - 2021 Intel Corporation */
#include <adf_accel_devices.h>
+#include <adf_admin.h>
#include <adf_clock.h>
#include <adf_common_drv.h>
#include <adf_gen2_config.h>
diff --git a/drivers/crypto/intel/qat/qat_c62x/adf_c62x_hw_data.c b/drivers/crypto/intel/qat/qat_c62x/adf_c62x_hw_data.c
index 355a781693eb..48cf3eb7c734 100644
--- a/drivers/crypto/intel/qat/qat_c62x/adf_c62x_hw_data.c
+++ b/drivers/crypto/intel/qat/qat_c62x/adf_c62x_hw_data.c
@@ -1,6 +1,7 @@
// SPDX-License-Identifier: (BSD-3-Clause OR GPL-2.0-only)
/* Copyright(c) 2014 - 2021 Intel Corporation */
#include <adf_accel_devices.h>
+#include <adf_admin.h>
#include <adf_clock.h>
#include <adf_common_drv.h>
#include <adf_gen2_config.h>
diff --git a/drivers/crypto/intel/qat/qat_common/adf_admin.c b/drivers/crypto/intel/qat/qat_common/adf_admin.c
index 3a04e743497f..15ffda582334 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_admin.c
+++ b/drivers/crypto/intel/qat/qat_common/adf_admin.c
@@ -7,6 +7,7 @@
#include <linux/pci.h>
#include <linux/dma-mapping.h>
#include "adf_accel_devices.h"
+#include "adf_admin.h"
#include "adf_common_drv.h"
#include "adf_cfg.h"
#include "adf_heartbeat.h"
diff --git a/drivers/crypto/intel/qat/qat_common/adf_admin.h b/drivers/crypto/intel/qat/qat_common/adf_admin.h
new file mode 100644
index 000000000000..03507ec3a51d
--- /dev/null
+++ b/drivers/crypto/intel/qat/qat_common/adf_admin.h
@@ -0,0 +1,19 @@
+/* SPDX-License-Identifier: GPL-2.0-only */
+/* Copyright(c) 2023 Intel Corporation */
+#ifndef ADF_ADMIN
+#define ADF_ADMIN
+
+struct adf_accel_dev;
+
+int adf_init_admin_comms(struct adf_accel_dev *accel_dev);
+void adf_exit_admin_comms(struct adf_accel_dev *accel_dev);
+int adf_send_admin_init(struct adf_accel_dev *accel_dev);
+int adf_get_ae_fw_counters(struct adf_accel_dev *accel_dev, u16 ae, u64 *reqs, u64 *resps);
+int adf_init_admin_pm(struct adf_accel_dev *accel_dev, u32 idle_delay);
+int adf_send_admin_tim_sync(struct adf_accel_dev *accel_dev, u32 cnt);
+int adf_send_admin_hb_timer(struct adf_accel_dev *accel_dev, uint32_t ticks);
+int adf_get_fw_timestamp(struct adf_accel_dev *accel_dev, u64 *timestamp);
+int adf_get_pm_info(struct adf_accel_dev *accel_dev, dma_addr_t p_state_addr, size_t buff_size);
+int adf_get_cnv_stats(struct adf_accel_dev *accel_dev, u16 ae, u16 *err_cnt, u16 *latest_err);
+
+#endif
diff --git a/drivers/crypto/intel/qat/qat_common/adf_clock.c b/drivers/crypto/intel/qat/qat_common/adf_clock.c
index dc0778691eb0..01e0a389e462 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_clock.c
+++ b/drivers/crypto/intel/qat/qat_common/adf_clock.c
@@ -10,6 +10,7 @@
#include <linux/types.h>
#include <linux/units.h>
#include <asm/errno.h>
+#include "adf_admin.h"
#include "adf_accel_devices.h"
#include "adf_clock.h"
#include "adf_common_drv.h"
diff --git a/drivers/crypto/intel/qat/qat_common/adf_cnv_dbgfs.c b/drivers/crypto/intel/qat/qat_common/adf_cnv_dbgfs.c
index aa5b6ff1dfb4..07119c487da0 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_cnv_dbgfs.c
+++ b/drivers/crypto/intel/qat/qat_common/adf_cnv_dbgfs.c
@@ -6,6 +6,7 @@
#include <linux/kernel.h>

#include "adf_accel_devices.h"
+#include "adf_admin.h"
#include "adf_common_drv.h"
#include "adf_cnv_dbgfs.h"
#include "qat_compression.h"
diff --git a/drivers/crypto/intel/qat/qat_common/adf_common_drv.h b/drivers/crypto/intel/qat/qat_common/adf_common_drv.h
index 18a382508542..ad9f4a464496 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_common_drv.h
+++ b/drivers/crypto/intel/qat/qat_common/adf_common_drv.h
@@ -87,16 +87,6 @@ void adf_reset_flr(struct adf_accel_dev *accel_dev);
void adf_dev_restore(struct adf_accel_dev *accel_dev);
int adf_init_aer(void);
void adf_exit_aer(void);
-int adf_init_admin_comms(struct adf_accel_dev *accel_dev);
-void adf_exit_admin_comms(struct adf_accel_dev *accel_dev);
-int adf_send_admin_init(struct adf_accel_dev *accel_dev);
-int adf_get_ae_fw_counters(struct adf_accel_dev *accel_dev, u16 ae, u64 *reqs, u64 *resps);
-int adf_init_admin_pm(struct adf_accel_dev *accel_dev, u32 idle_delay);
-int adf_send_admin_tim_sync(struct adf_accel_dev *accel_dev, u32 cnt);
-int adf_send_admin_hb_timer(struct adf_accel_dev *accel_dev, uint32_t ticks);
-int adf_get_fw_timestamp(struct adf_accel_dev *accel_dev, u64 *timestamp);
-int adf_get_pm_info(struct adf_accel_dev *accel_dev, dma_addr_t p_state_addr, size_t buff_size);
-int adf_get_cnv_stats(struct adf_accel_dev *accel_dev, u16 ae, u16 *err_cnt, u16 *latest_err);
int adf_init_arb(struct adf_accel_dev *accel_dev);
void adf_exit_arb(struct adf_accel_dev *accel_dev);
void adf_update_ring_arb(struct adf_etr_ring_data *ring);
diff --git a/drivers/crypto/intel/qat/qat_common/adf_fw_counters.c b/drivers/crypto/intel/qat/qat_common/adf_fw_counters.c
index 6abe4736eab8..98fb7ccfed9f 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_fw_counters.c
+++ b/drivers/crypto/intel/qat/qat_common/adf_fw_counters.c
@@ -9,6 +9,7 @@
#include <linux/types.h>

#include "adf_accel_devices.h"
+#include "adf_admin.h"
#include "adf_common_drv.h"
#include "adf_fw_counters.h"

diff --git a/drivers/crypto/intel/qat/qat_common/adf_gen4_pm.c b/drivers/crypto/intel/qat/qat_common/adf_gen4_pm.c
index c663d3a20c5b..5dafd9a270db 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_gen4_pm.c
+++ b/drivers/crypto/intel/qat/qat_common/adf_gen4_pm.c
@@ -5,6 +5,7 @@
#include <linux/kernel.h>

#include "adf_accel_devices.h"
+#include "adf_admin.h"
#include "adf_common_drv.h"
#include "adf_gen4_pm.h"
#include "adf_cfg_strings.h"
diff --git a/drivers/crypto/intel/qat/qat_common/adf_gen4_pm_debugfs.c b/drivers/crypto/intel/qat/qat_common/adf_gen4_pm_debugfs.c
index 5114759287c6..ee0b5079de3e 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_gen4_pm_debugfs.c
+++ b/drivers/crypto/intel/qat/qat_common/adf_gen4_pm_debugfs.c
@@ -6,6 +6,7 @@
#include <linux/stringify.h>

#include "adf_accel_devices.h"
+#include "adf_admin.h"
#include "adf_common_drv.h"
#include "adf_gen4_pm.h"
#include "icp_qat_fw_init_admin.h"
diff --git a/drivers/crypto/intel/qat/qat_common/adf_gen4_timer.c b/drivers/crypto/intel/qat/qat_common/adf_gen4_timer.c
index 646c57922fcd..35ccb91d6ec1 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_gen4_timer.c
+++ b/drivers/crypto/intel/qat/qat_common/adf_gen4_timer.c
@@ -9,6 +9,7 @@
#include <linux/slab.h>
#include <linux/workqueue.h>

+#include "adf_admin.h"
#include "adf_accel_devices.h"
#include "adf_common_drv.h"
#include "adf_gen4_timer.h"
diff --git a/drivers/crypto/intel/qat/qat_common/adf_heartbeat.c b/drivers/crypto/intel/qat/qat_common/adf_heartbeat.c
index beef9a5f6c75..13f48d2f6da8 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_heartbeat.c
+++ b/drivers/crypto/intel/qat/qat_common/adf_heartbeat.c
@@ -12,6 +12,7 @@
#include <linux/types.h>
#include <asm/errno.h>
#include "adf_accel_devices.h"
+#include "adf_admin.h"
#include "adf_cfg.h"
#include "adf_cfg_strings.h"
#include "adf_clock.h"
diff --git a/drivers/crypto/intel/qat/qat_common/adf_heartbeat_dbgfs.c b/drivers/crypto/intel/qat/qat_common/adf_heartbeat_dbgfs.c
index 803cbfd838f0..2661af6a2ef6 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_heartbeat_dbgfs.c
+++ b/drivers/crypto/intel/qat/qat_common/adf_heartbeat_dbgfs.c
@@ -8,6 +8,7 @@
#include <linux/kernel.h>
#include <linux/kstrtox.h>
#include <linux/types.h>
+#include "adf_admin.h"
#include "adf_cfg.h"
#include "adf_common_drv.h"
#include "adf_heartbeat.h"
diff --git a/drivers/crypto/intel/qat/qat_dh895xcc/adf_dh895xcc_hw_data.c b/drivers/crypto/intel/qat/qat_dh895xcc/adf_dh895xcc_hw_data.c
index 09551f949126..af14090cc4be 100644
--- a/drivers/crypto/intel/qat/qat_dh895xcc/adf_dh895xcc_hw_data.c
+++ b/drivers/crypto/intel/qat/qat_dh895xcc/adf_dh895xcc_hw_data.c
@@ -1,6 +1,7 @@
// SPDX-License-Identifier: (BSD-3-Clause OR GPL-2.0-only)
/* Copyright(c) 2014 - 2021 Intel Corporation */
#include <adf_accel_devices.h>
+#include <adf_admin.h>
#include <adf_common_drv.h>
#include <adf_gen2_config.h>
#include <adf_gen2_dc.h>
--
2.41.0

2023-10-11 12:36:30

by Damian Muszynski

[permalink] [raw]
Subject: [PATCH 03/11] crypto: qat - fix ring to service map for QAT GEN4

From: Giovanni Cabiddu <[email protected]>

The 4xxx drivers hardcode the ring to service mapping. However, when
additional configurations where added to the driver, the mappings were
not updated. This implies that an incorrect mapping might be reported
through pfvf for certain configurations.

Add an algorithm that computes the correct ring to service mapping based
on the firmware loaded on the device.

Fixes: 0cec19c761e5 ("crypto: qat - add support for compression for 4xxx")
Signed-off-by: Giovanni Cabiddu <[email protected]>
Reviewed-by: Damian Muszynski <[email protected]>
---
.../intel/qat/qat_4xxx/adf_4xxx_hw_data.c | 54 +++++++++++++++++++
.../intel/qat/qat_common/adf_accel_devices.h | 1 +
.../crypto/intel/qat/qat_common/adf_init.c | 3 ++
3 files changed, 58 insertions(+)

diff --git a/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c b/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
index 44b732fb80bc..a5691ba0b724 100644
--- a/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
+++ b/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
@@ -423,6 +423,59 @@ static const struct adf_fw_config *get_fw_config(struct adf_accel_dev *accel_dev
}
}

+enum adf_rp_groups {
+ RP_GROUP_0 = 0,
+ RP_GROUP_1,
+ RP_GROUP_COUNT
+};
+
+static u16 get_ring_to_svc_map(struct adf_accel_dev *accel_dev)
+{
+ enum adf_cfg_service_type rps[RP_GROUP_COUNT];
+ const struct adf_fw_config *fw_config;
+ u16 ring_to_svc_map;
+ int i, j;
+
+ fw_config = get_fw_config(accel_dev);
+ if (!fw_config)
+ return 0;
+
+ for (i = 0; i < RP_GROUP_COUNT; i++) {
+ switch (fw_config[i].ae_mask) {
+ case ADF_AE_GROUP_0:
+ j = RP_GROUP_0;
+ break;
+ case ADF_AE_GROUP_1:
+ j = RP_GROUP_1;
+ break;
+ default:
+ return 0;
+ }
+
+ switch (fw_config[i].obj) {
+ case ADF_FW_SYM_OBJ:
+ rps[j] = SYM;
+ break;
+ case ADF_FW_ASYM_OBJ:
+ rps[j] = ASYM;
+ break;
+ case ADF_FW_DC_OBJ:
+ rps[j] = COMP;
+ break;
+ default:
+ rps[j] = 0;
+ break;
+ }
+ }
+
+ ring_to_svc_map = rps[RP_GROUP_0] << ADF_CFG_SERV_RING_PAIR_0_SHIFT |
+ rps[RP_GROUP_1] << ADF_CFG_SERV_RING_PAIR_1_SHIFT |
+ rps[RP_GROUP_0] << ADF_CFG_SERV_RING_PAIR_2_SHIFT |
+ rps[RP_GROUP_1] << ADF_CFG_SERV_RING_PAIR_3_SHIFT;
+
+ return ring_to_svc_map;
+}
+
static const char *uof_get_name(struct adf_accel_dev *accel_dev, u32 obj_num,
const char * const fw_objs[], int num_objs)
{
@@ -519,6 +572,7 @@ void adf_init_hw_data_4xxx(struct adf_hw_device_data *hw_data, u32 dev_id)
hw_data->uof_get_ae_mask = uof_get_ae_mask;
hw_data->set_msix_rttable = set_msix_default_rttable;
hw_data->set_ssm_wdtimer = adf_gen4_set_ssm_wdtimer;
+ hw_data->get_ring_to_svc_map = get_ring_to_svc_map;
hw_data->disable_iov = adf_disable_sriov;
hw_data->ring_pair_reset = adf_gen4_ring_pair_reset;
hw_data->enable_pm = adf_gen4_enable_pm;
diff --git a/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h b/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
index 9677c8e0f180..3674904d0527 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
+++ b/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
@@ -182,6 +182,7 @@ struct adf_hw_device_data {
void (*get_arb_info)(struct arb_info *arb_csrs_info);
void (*get_admin_info)(struct admin_info *admin_csrs_info);
enum dev_sku_info (*get_sku)(struct adf_hw_device_data *self);
+ u16 (*get_ring_to_svc_map)(struct adf_accel_dev *accel_dev);
int (*alloc_irq)(struct adf_accel_dev *accel_dev);
void (*free_irq)(struct adf_accel_dev *accel_dev);
void (*enable_error_correction)(struct adf_accel_dev *accel_dev);
diff --git a/drivers/crypto/intel/qat/qat_common/adf_init.c b/drivers/crypto/intel/qat/qat_common/adf_init.c
index bccd6bf8cf63..b4cf605ccf3e 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_init.c
+++ b/drivers/crypto/intel/qat/qat_common/adf_init.c
@@ -96,6 +96,9 @@ static int adf_dev_init(struct adf_accel_dev *accel_dev)
return -EFAULT;
}

+ if (hw_data->get_ring_to_svc_map)
+ hw_data->ring_to_svc_map = hw_data->get_ring_to_svc_map(accel_dev);
+
if (adf_ae_init(accel_dev)) {
dev_err(&GET_DEV(accel_dev),
"Failed to initialise Acceleration Engine\n");
--
2.41.0

2023-10-11 12:36:33

by Damian Muszynski

[permalink] [raw]
Subject: [PATCH 05/11] units: Add BYTES_PER_*BIT

There is going to be a new user of the BYTES_PER_[K/M/G]BIT definition
besides possibly existing ones. Add them to the header.

Signed-off-by: Damian Muszynski <[email protected]>
Reviewed-by: Giovanni Cabiddu <[email protected]>
---
include/linux/units.h | 4 ++++
1 file changed, 4 insertions(+)

diff --git a/include/linux/units.h b/include/linux/units.h
index 2793a41e73a2..ff1bd6b5f5b3 100644
--- a/include/linux/units.h
+++ b/include/linux/units.h
@@ -31,6 +31,10 @@
#define MICROWATT_PER_MILLIWATT 1000UL
#define MICROWATT_PER_WATT 1000000UL

+#define BYTES_PER_KBIT (KILO / BITS_PER_BYTE)
+#define BYTES_PER_MBIT (MEGA / BITS_PER_BYTE)
+#define BYTES_PER_GBIT (GIGA / BITS_PER_BYTE)
+
#define ABSOLUTE_ZERO_MILLICELSIUS -273150

static inline long milli_kelvin_to_millicelsius(long t)
--
2.41.0

2023-10-11 12:36:33

by Damian Muszynski

[permalink] [raw]
Subject: [PATCH 07/11] crypto: qat - add retrieval of fw capabilities

The QAT firmware provides a mechanism to retrieve its capabilities
through the init admin interface.

Add logic to retrieve the firmware capability mask from the firmware
through the init/admin channel. This mask reports if the
power management, telemetry and rate limiting features are supported.

The fw capabilities are stored in the accel_dev structure and are used
to detect if a certain feature is supported by the firmware loaded
in the device.

This is supported only by devices which have an admin AE.

Signed-off-by: Damian Muszynski <[email protected]>
Reviewed-by: Giovanni Cabiddu <[email protected]>
---
.../intel/qat/qat_common/adf_accel_devices.h | 1 +
.../crypto/intel/qat/qat_common/adf_admin.c | 25 +++++++++++++++++++
.../qat/qat_common/icp_qat_fw_init_admin.h | 3 +++
3 files changed, 29 insertions(+)

diff --git a/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h b/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
index 3674904d0527..45742226a96f 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
+++ b/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
@@ -221,6 +221,7 @@ struct adf_hw_device_data {
u32 straps;
u32 accel_capabilities_mask;
u32 extended_dc_capabilities;
+ u16 fw_capabilities;
u32 clock_frequency;
u32 instance_id;
u16 accel_mask;
diff --git a/drivers/crypto/intel/qat/qat_common/adf_admin.c b/drivers/crypto/intel/qat/qat_common/adf_admin.c
index 15ffda582334..9ff00eb4cc67 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_admin.c
+++ b/drivers/crypto/intel/qat/qat_common/adf_admin.c
@@ -310,6 +310,26 @@ static bool is_dcc_enabled(struct adf_accel_dev *accel_dev)
return !strcmp(services, "dcc");
}

+static int adf_get_fw_capabilities(struct adf_accel_dev *accel_dev, u16 *caps)
+{
+ u32 ae_mask = accel_dev->hw_device->admin_ae_mask;
+ struct icp_qat_fw_init_admin_resp resp = { };
+ struct icp_qat_fw_init_admin_req req = { };
+ int ret;
+
+ if (!ae_mask)
+ return 0;
+
+ req.cmd_id = ICP_QAT_FW_CAPABILITIES_GET;
+ ret = adf_send_admin(accel_dev, &req, &resp, ae_mask);
+ if (ret)
+ return ret;
+
+ *caps = resp.fw_capabilities;
+
+ return 0;
+}
+
/**
* adf_send_admin_init() - Function sends init message to FW
* @accel_dev: Pointer to acceleration device.
@@ -321,6 +341,7 @@ static bool is_dcc_enabled(struct adf_accel_dev *accel_dev)
int adf_send_admin_init(struct adf_accel_dev *accel_dev)
{
u32 dc_capabilities = 0;
+ u16 fw_capabilities = 0;
int ret;

ret = adf_set_fw_constants(accel_dev);
@@ -340,6 +361,10 @@ int adf_send_admin_init(struct adf_accel_dev *accel_dev)
}
accel_dev->hw_device->extended_dc_capabilities = dc_capabilities;

+ ret = adf_get_fw_capabilities(accel_dev, &fw_capabilities);
+ if (!ret)
+ accel_dev->hw_device->fw_capabilities = fw_capabilities;
+
return adf_init_ae(accel_dev);
}
EXPORT_SYMBOL_GPL(adf_send_admin_init);
diff --git a/drivers/crypto/intel/qat/qat_common/icp_qat_fw_init_admin.h b/drivers/crypto/intel/qat/qat_common/icp_qat_fw_init_admin.h
index 9e5ce419d875..e4de9a30e0bd 100644
--- a/drivers/crypto/intel/qat/qat_common/icp_qat_fw_init_admin.h
+++ b/drivers/crypto/intel/qat/qat_common/icp_qat_fw_init_admin.h
@@ -16,6 +16,7 @@ enum icp_qat_fw_init_admin_cmd_id {
ICP_QAT_FW_HEARTBEAT_SYNC = 7,
ICP_QAT_FW_HEARTBEAT_GET = 8,
ICP_QAT_FW_COMP_CAPABILITY_GET = 9,
+ ICP_QAT_FW_CRYPTO_CAPABILITY_GET = 10,
ICP_QAT_FW_DC_CHAIN_INIT = 11,
ICP_QAT_FW_HEARTBEAT_TIMER_SET = 13,
ICP_QAT_FW_TIMER_GET = 19,
@@ -109,10 +110,12 @@ struct icp_qat_fw_init_admin_resp {
__u32 unsuccessful_count;
__u64 resrvd8;
};
+ __u16 fw_capabilities;
};
} __packed;

#define ICP_QAT_FW_SYNC ICP_QAT_FW_HEARTBEAT_SYNC
+#define ICP_QAT_FW_CAPABILITIES_GET ICP_QAT_FW_CRYPTO_CAPABILITY_GET

#define ICP_QAT_NUMBER_OF_PM_EVENTS 8

--
2.41.0

2023-10-11 12:36:36

by Damian Muszynski

[permalink] [raw]
Subject: [PATCH 06/11] crypto: qat - add bits.h to icp_qat_hw.h

Some enums use the macro BIT. Include bits.h as it is missing.

Signed-off-by: Damian Muszynski <[email protected]>
Reviewed-by: Giovanni Cabiddu <[email protected]>
---
drivers/crypto/intel/qat/qat_common/icp_qat_hw.h | 2 ++
1 file changed, 2 insertions(+)

diff --git a/drivers/crypto/intel/qat/qat_common/icp_qat_hw.h b/drivers/crypto/intel/qat/qat_common/icp_qat_hw.h
index 0c8883e2ccc6..eb2ef225bcee 100644
--- a/drivers/crypto/intel/qat/qat_common/icp_qat_hw.h
+++ b/drivers/crypto/intel/qat/qat_common/icp_qat_hw.h
@@ -3,6 +3,8 @@
#ifndef _ICP_QAT_HW_H_
#define _ICP_QAT_HW_H_

+#include <linux/bits.h>
+
enum icp_qat_hw_ae_id {
ICP_QAT_HW_AE_0 = 0,
ICP_QAT_HW_AE_1 = 1,
--
2.41.0

2023-10-11 12:36:44

by Damian Muszynski

[permalink] [raw]
Subject: [PATCH 08/11] crypto: qat - add rate limiting feature to qat_4xxx

The Rate Limiting (RL) feature allows to control the rate of requests
that can be submitted on a ring pair (RP). This allows sharing a QAT
device among multiple users while ensuring a guaranteed throughput.

Configuration of RL is accomplished through entities called SLAs
(Service Level Agreement). Each SLA object gets a unique identifier
and defines the limitations for a single service across up to four
ring pairs (RPs count allocated to a single VF).

The rate is determined using two fields:
* CIR (Committed Information Rate), i.e., the guaranteed rate.
* PIR (Peak Information Rate), i.e., the maximum rate achievable
when the device has available resources.
The rate values are expressed in permille scale i.e. 0-1000.
Ring pair selection is achieved by providing a 64-bit mask, where
each bit corresponds to one of the ring pairs.

This adds an interface and logic that allow to add, update, retrieve
and remove an SLA.

Signed-off-by: Damian Muszynski <[email protected]>
Reviewed-by: Giovanni Cabiddu <[email protected]>
---
.../intel/qat/qat_4xxx/adf_4xxx_hw_data.c | 20 +
.../intel/qat/qat_4xxx/adf_4xxx_hw_data.h | 13 +-
drivers/crypto/intel/qat/qat_common/Makefile | 2 +
.../intel/qat/qat_common/adf_accel_devices.h | 3 +
.../crypto/intel/qat/qat_common/adf_admin.c | 47 +
.../crypto/intel/qat/qat_common/adf_admin.h | 8 +
.../intel/qat/qat_common/adf_gen4_hw_data.h | 7 +
.../crypto/intel/qat/qat_common/adf_init.c | 9 +
drivers/crypto/intel/qat/qat_common/adf_rl.c | 1176 +++++++++++++++++
drivers/crypto/intel/qat/qat_common/adf_rl.h | 169 +++
.../intel/qat/qat_common/adf_rl_admin.c | 98 ++
.../intel/qat/qat_common/adf_rl_admin.h | 18 +
.../qat/qat_common/icp_qat_fw_init_admin.h | 38 +
13 files changed, 1607 insertions(+), 1 deletion(-)
create mode 100644 drivers/crypto/intel/qat/qat_common/adf_rl.c
create mode 100644 drivers/crypto/intel/qat/qat_common/adf_rl.h
create mode 100644 drivers/crypto/intel/qat/qat_common/adf_rl_admin.c
create mode 100644 drivers/crypto/intel/qat/qat_common/adf_rl_admin.h

diff --git a/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c b/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
index 8a80701b7791..bc3745f0018f 100644
--- a/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
+++ b/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
@@ -342,6 +342,24 @@ static u32 get_heartbeat_clock(struct adf_hw_device_data *self)
return ADF_4XXX_KPT_COUNTER_FREQ;
}

+static void adf_init_rl_data(struct adf_rl_hw_data *rl_data)
+{
+ rl_data->pciout_tb_offset = ADF_GEN4_RL_TOKEN_PCIEOUT_BUCKET_OFFSET;
+ rl_data->pciin_tb_offset = ADF_GEN4_RL_TOKEN_PCIEIN_BUCKET_OFFSET;
+ rl_data->r2l_offset = ADF_GEN4_RL_R2L_OFFSET;
+ rl_data->l2c_offset = ADF_GEN4_RL_L2C_OFFSET;
+ rl_data->c2s_offset = ADF_GEN4_RL_C2S_OFFSET;
+
+ rl_data->pcie_scale_div = ADF_4XXX_RL_PCIE_SCALE_FACTOR_DIV;
+ rl_data->pcie_scale_mul = ADF_4XXX_RL_PCIE_SCALE_FACTOR_MUL;
+ rl_data->dcpr_correction = ADF_4XXX_RL_DCPR_CORRECTION;
+ rl_data->max_tp[ADF_SVC_ASYM] = ADF_4XXX_RL_MAX_TP_ASYM;
+ rl_data->max_tp[ADF_SVC_SYM] = ADF_4XXX_RL_MAX_TP_SYM;
+ rl_data->max_tp[ADF_SVC_DC] = ADF_4XXX_RL_MAX_TP_DC;
+ rl_data->scan_interval = ADF_4XXX_RL_SCANS_PER_SEC;
+ rl_data->scale_ref = ADF_4XXX_RL_SLICE_REF;
+}
+
static void adf_enable_error_correction(struct adf_accel_dev *accel_dev)
{
struct adf_bar *misc_bar = &GET_BARS(accel_dev)[ADF_4XXX_PMISC_BAR];
@@ -583,10 +601,12 @@ void adf_init_hw_data_4xxx(struct adf_hw_device_data *hw_data, u32 dev_id)
hw_data->stop_timer = adf_gen4_timer_stop;
hw_data->get_hb_clock = get_heartbeat_clock;
hw_data->num_hb_ctrs = ADF_NUM_HB_CNT_PER_AE;
+ hw_data->clock_frequency = ADF_4XXX_AE_FREQ;

adf_gen4_init_hw_csr_ops(&hw_data->csr_ops);
adf_gen4_init_pf_pfvf_ops(&hw_data->pfvf_ops);
adf_gen4_init_dc_ops(&hw_data->dc_ops);
+ adf_init_rl_data(&hw_data->rl_data);
}

void adf_clean_hw_data_4xxx(struct adf_hw_device_data *hw_data)
diff --git a/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.h b/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.h
index bb3d95a8fb21..590cbe816b7b 100644
--- a/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.h
+++ b/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.h
@@ -65,8 +65,19 @@
#define ADF_402XX_ASYM_OBJ "qat_402xx_asym.bin"
#define ADF_402XX_ADMIN_OBJ "qat_402xx_admin.bin"

+/* RL constants */
+#define ADF_4XXX_RL_PCIE_SCALE_FACTOR_DIV 100
+#define ADF_4XXX_RL_PCIE_SCALE_FACTOR_MUL 102
+#define ADF_4XXX_RL_DCPR_CORRECTION 1
+#define ADF_4XXX_RL_SCANS_PER_SEC 954
+#define ADF_4XXX_RL_MAX_TP_ASYM 173750UL
+#define ADF_4XXX_RL_MAX_TP_SYM 95000UL
+#define ADF_4XXX_RL_MAX_TP_DC 45000UL
+#define ADF_4XXX_RL_SLICE_REF 1000UL
+
/* Clocks frequency */
-#define ADF_4XXX_KPT_COUNTER_FREQ (100 * HZ_PER_MHZ)
+#define ADF_4XXX_KPT_COUNTER_FREQ (100 * HZ_PER_MHZ)
+#define ADF_4XXX_AE_FREQ (1000 * HZ_PER_MHZ)

/* qat_4xxx fuse bits are different from old GENs, redefine them */
enum icp_qat_4xxx_slice_mask {
diff --git a/drivers/crypto/intel/qat/qat_common/Makefile b/drivers/crypto/intel/qat/qat_common/Makefile
index 204c7d0aa31e..77f8c484d49c 100644
--- a/drivers/crypto/intel/qat/qat_common/Makefile
+++ b/drivers/crypto/intel/qat/qat_common/Makefile
@@ -26,6 +26,8 @@ intel_qat-objs := adf_cfg.o \
qat_algs.o \
qat_asym_algs.o \
qat_algs_send.o \
+ adf_rl.o \
+ adf_rl_admin.o \
qat_uclo.o \
qat_hal.o \
qat_bl.o
diff --git a/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h b/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
index 45742226a96f..35b805b9a136 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
+++ b/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
@@ -8,6 +8,7 @@
#include <linux/io.h>
#include <linux/ratelimit.h>
#include "adf_cfg_common.h"
+#include "adf_rl.h"
#include "adf_pfvf_msg.h"

#define ADF_DH895XCC_DEVICE_NAME "dh895xcc"
@@ -215,6 +216,7 @@ struct adf_hw_device_data {
struct adf_pfvf_ops pfvf_ops;
struct adf_hw_csr_ops csr_ops;
struct adf_dc_ops dc_ops;
+ struct adf_rl_hw_data rl_data;
const char *fw_name;
const char *fw_mmp_name;
u32 fuses;
@@ -325,6 +327,7 @@ struct adf_accel_dev {
struct adf_accel_pci accel_pci_dev;
struct adf_timer *timer;
struct adf_heartbeat *heartbeat;
+ struct adf_rl *rate_limiting;
union {
struct {
/* protects VF2PF interrupts access */
diff --git a/drivers/crypto/intel/qat/qat_common/adf_admin.c b/drivers/crypto/intel/qat/qat_common/adf_admin.c
index 9ff00eb4cc67..b9cd21f772d3 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_admin.c
+++ b/drivers/crypto/intel/qat/qat_common/adf_admin.c
@@ -330,6 +330,53 @@ static int adf_get_fw_capabilities(struct adf_accel_dev *accel_dev, u16 *caps)
return 0;
}

+int adf_send_admin_rl_init(struct adf_accel_dev *accel_dev,
+ struct icp_qat_fw_init_admin_slice_cnt *slices)
+{
+ u32 ae_mask = accel_dev->hw_device->admin_ae_mask;
+ struct icp_qat_fw_init_admin_resp resp = { };
+ struct icp_qat_fw_init_admin_req req = { };
+ int ret;
+
+ req.cmd_id = ICP_QAT_FW_RL_INIT;
+
+ ret = adf_send_admin(accel_dev, &req, &resp, ae_mask);
+ if (ret)
+ return ret;
+
+ memcpy(slices, &resp.slices, sizeof(*slices));
+
+ return 0;
+}
+
+int adf_send_admin_rl_add_update(struct adf_accel_dev *accel_dev,
+ struct icp_qat_fw_init_admin_req *req)
+{
+ u32 ae_mask = accel_dev->hw_device->admin_ae_mask;
+ struct icp_qat_fw_init_admin_resp resp = { };
+
+ /*
+ * req struct filled in rl implementation. Used commands
+ * ICP_QAT_FW_RL_ADD for a new SLA
+ * ICP_QAT_FW_RL_UPDATE for update SLA
+ */
+ return adf_send_admin(accel_dev, req, &resp, ae_mask);
+}
+
+int adf_send_admin_rl_delete(struct adf_accel_dev *accel_dev, u16 node_id,
+ u8 node_type)
+{
+ u32 ae_mask = accel_dev->hw_device->admin_ae_mask;
+ struct icp_qat_fw_init_admin_resp resp = { };
+ struct icp_qat_fw_init_admin_req req = { };
+
+ req.cmd_id = ICP_QAT_FW_RL_REMOVE;
+ req.node_id = node_id;
+ req.node_type = node_type;
+
+ return adf_send_admin(accel_dev, &req, &resp, ae_mask);
+}
+
/**
* adf_send_admin_init() - Function sends init message to FW
* @accel_dev: Pointer to acceleration device.
diff --git a/drivers/crypto/intel/qat/qat_common/adf_admin.h b/drivers/crypto/intel/qat/qat_common/adf_admin.h
index 03507ec3a51d..55cbcbc66c9f 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_admin.h
+++ b/drivers/crypto/intel/qat/qat_common/adf_admin.h
@@ -3,6 +3,8 @@
#ifndef ADF_ADMIN
#define ADF_ADMIN

+#include "icp_qat_fw_init_admin.h"
+
struct adf_accel_dev;

int adf_init_admin_comms(struct adf_accel_dev *accel_dev);
@@ -12,6 +14,12 @@ int adf_get_ae_fw_counters(struct adf_accel_dev *accel_dev, u16 ae, u64 *reqs, u
int adf_init_admin_pm(struct adf_accel_dev *accel_dev, u32 idle_delay);
int adf_send_admin_tim_sync(struct adf_accel_dev *accel_dev, u32 cnt);
int adf_send_admin_hb_timer(struct adf_accel_dev *accel_dev, uint32_t ticks);
+int adf_send_admin_rl_init(struct adf_accel_dev *accel_dev,
+ struct icp_qat_fw_init_admin_slice_cnt *slices);
+int adf_send_admin_rl_add_update(struct adf_accel_dev *accel_dev,
+ struct icp_qat_fw_init_admin_req *req);
+int adf_send_admin_rl_delete(struct adf_accel_dev *accel_dev, u16 node_id,
+ u8 node_type);
int adf_get_fw_timestamp(struct adf_accel_dev *accel_dev, u64 *timestamp);
int adf_get_pm_info(struct adf_accel_dev *accel_dev, dma_addr_t p_state_addr, size_t buff_size);
int adf_get_cnv_stats(struct adf_accel_dev *accel_dev, u16 ae, u16 *err_cnt, u16 *latest_err);
diff --git a/drivers/crypto/intel/qat/qat_common/adf_gen4_hw_data.h b/drivers/crypto/intel/qat/qat_common/adf_gen4_hw_data.h
index 02d7a019ebf8..1813fe1d5a06 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_gen4_hw_data.h
+++ b/drivers/crypto/intel/qat/qat_common/adf_gen4_hw_data.h
@@ -139,6 +139,13 @@ do { \
/* Number of heartbeat counter pairs */
#define ADF_NUM_HB_CNT_PER_AE ADF_NUM_THREADS_PER_AE

+/* Rate Limiting */
+#define ADF_GEN4_RL_R2L_OFFSET 0x508000
+#define ADF_GEN4_RL_L2C_OFFSET 0x509000
+#define ADF_GEN4_RL_C2S_OFFSET 0x508818
+#define ADF_GEN4_RL_TOKEN_PCIEIN_BUCKET_OFFSET 0x508800
+#define ADF_GEN4_RL_TOKEN_PCIEOUT_BUCKET_OFFSET 0x508804
+
void adf_gen4_set_ssm_wdtimer(struct adf_accel_dev *accel_dev);
void adf_gen4_init_hw_csr_ops(struct adf_hw_csr_ops *csr_ops);
int adf_gen4_ring_pair_reset(struct adf_accel_dev *accel_dev, u32 bank_number);
diff --git a/drivers/crypto/intel/qat/qat_common/adf_init.c b/drivers/crypto/intel/qat/qat_common/adf_init.c
index b4cf605ccf3e..827c70c4221f 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_init.c
+++ b/drivers/crypto/intel/qat/qat_common/adf_init.c
@@ -9,6 +9,7 @@
#include "adf_common_drv.h"
#include "adf_dbgfs.h"
#include "adf_heartbeat.h"
+#include "adf_rl.h"

static LIST_HEAD(service_table);
static DEFINE_MUTEX(service_lock);
@@ -133,6 +134,9 @@ static int adf_dev_init(struct adf_accel_dev *accel_dev)
}

adf_heartbeat_init(accel_dev);
+ ret = adf_rl_init(accel_dev);
+ if (ret && ret != -EOPNOTSUPP)
+ return ret;

/*
* Subservice initialisation is divided into two stages: init and start.
@@ -208,6 +212,9 @@ static int adf_dev_start(struct adf_accel_dev *accel_dev)
}

adf_heartbeat_start(accel_dev);
+ ret = adf_rl_start(accel_dev);
+ if (ret && ret != -EOPNOTSUPP)
+ return ret;

list_for_each_entry(service, &service_table, list) {
if (service->event_hld(accel_dev, ADF_EVENT_START)) {
@@ -267,6 +274,7 @@ static void adf_dev_stop(struct adf_accel_dev *accel_dev)
!test_bit(ADF_STATUS_STARTING, &accel_dev->status))
return;

+ adf_rl_stop(accel_dev);
adf_dbgfs_rm(accel_dev);

clear_bit(ADF_STATUS_STARTING, &accel_dev->status);
@@ -353,6 +361,7 @@ static void adf_dev_shutdown(struct adf_accel_dev *accel_dev)
clear_bit(accel_dev->accel_id, service->init_status);
}

+ adf_rl_exit(accel_dev);
adf_heartbeat_shutdown(accel_dev);

hw_data->disable_iov(accel_dev);
diff --git a/drivers/crypto/intel/qat/qat_common/adf_rl.c b/drivers/crypto/intel/qat/qat_common/adf_rl.c
new file mode 100644
index 000000000000..dec98a056dce
--- /dev/null
+++ b/drivers/crypto/intel/qat/qat_common/adf_rl.c
@@ -0,0 +1,1176 @@
+// SPDX-License-Identifier: GPL-2.0-only
+/* Copyright(c) 2023 Intel Corporation */
+
+#define dev_fmt(fmt) "RateLimiting: " fmt
+
+#include <asm/errno.h>
+#include <asm/div64.h>
+
+#include <linux/dev_printk.h>
+#include <linux/kernel.h>
+#include <linux/pci.h>
+#include <linux/slab.h>
+#include <linux/units.h>
+
+#include "adf_accel_devices.h"
+#include "adf_common_drv.h"
+#include "adf_rl_admin.h"
+#include "adf_rl.h"
+
+#define RL_TOKEN_GRANULARITY_PCIEIN_BUCKET 0U
+#define RL_TOKEN_GRANULARITY_PCIEOUT_BUCKET 0U
+#define RL_TOKEN_PCIE_SIZE 64
+#define RL_TOKEN_ASYM_SIZE 1024
+#define RL_CSR_SIZE 4U
+#define RL_CAPABILITY_MASK GENMASK(6, 4)
+#define RL_CAPABILITY_VALUE 0x70
+#define RL_VALIDATE_NON_ZERO(input) ((input) == 0)
+#define ROOT_MASK GENMASK(1, 0)
+#define CLUSTER_MASK GENMASK(3, 0)
+#define LEAF_MASK GENMASK(5, 0)
+
+static int validate_user_input(struct adf_accel_dev *accel_dev,
+ struct adf_rl_sla_input_data *sla_in,
+ bool is_update)
+{
+ const unsigned long rp_mask = sla_in->rp_mask;
+ size_t rp_mask_size;
+ int i, cnt;
+
+ if (sla_in->sla_id < RL_SLA_EMPTY_ID || sla_in->sla_id >= RL_NODES_CNT_MAX) {
+ dev_notice(&GET_DEV(accel_dev), "Wrong SLA ID\n");
+ goto ret_inval;
+ }
+
+ if (sla_in->pir < sla_in->cir) {
+ dev_notice(&GET_DEV(accel_dev),
+ "PIR must be >= CIR, setting PIR to CIR\n");
+ sla_in->pir = sla_in->cir;
+ }
+
+ if (!is_update) {
+ cnt = 0;
+ rp_mask_size = sizeof(sla_in->rp_mask) * BITS_PER_BYTE;
+ for_each_set_bit(i, &rp_mask, rp_mask_size) {
+ if (++cnt > RL_RP_CNT_PER_LEAF_MAX) {
+ dev_notice(&GET_DEV(accel_dev),
+ "Too many ring pairs selected for this SLA\n");
+ goto ret_inval;
+ }
+ }
+
+ if (sla_in->srv >= ADF_SVC_NONE) {
+ dev_notice(&GET_DEV(accel_dev),
+ "Wrong service type\n");
+ goto ret_inval;
+ }
+
+ if (sla_in->type > RL_LEAF) {
+ dev_notice(&GET_DEV(accel_dev),
+ "Wrong node type\n");
+ goto ret_inval;
+ }
+
+ if (sla_in->parent_id < RL_PARENT_DEFAULT_ID ||
+ sla_in->parent_id >= RL_NODES_CNT_MAX) {
+ dev_notice(&GET_DEV(accel_dev),
+ "Wrong parent ID\n");
+ goto ret_inval;
+ }
+ }
+
+ return 0;
+
+ret_inval:
+ return -EINVAL;
+}
+
+static int validate_sla_id(struct adf_accel_dev *accel_dev, int sla_id)
+{
+ struct rl_sla *sla;
+
+ if (sla_id < 0 || sla_id >= RL_NODES_CNT_MAX)
+ goto ret_not_exists;
+
+ sla = accel_dev->rate_limiting->sla[sla_id];
+
+ if (!sla)
+ goto ret_not_exists;
+
+ if (sla->type != RL_LEAF) {
+ dev_notice(&GET_DEV(accel_dev),
+ "This ID is reserved for internal use\n");
+ goto ret_inval;
+ }
+
+ return 0;
+
+ret_not_exists:
+ dev_notice(&GET_DEV(accel_dev),
+ "SLA with provided ID does not exist\n");
+ret_inval:
+ return -EINVAL;
+}
+
+/**
+ * find_parent() - Find the parent for a new SLA
+ * @rl_data: pointer to ratelimiting data
+ * @sla_in: pointer to user input data for a new SLA
+ *
+ * Function returns a pointer to the parent SLA. If the parent ID is provided
+ * as input in the user data, then such ID is validated and the parent SLA
+ * is returned.
+ * Otherwise, it returns the default parent SLA (root or cluster) for
+ * the new object.
+ *
+ * Return:
+ * * Pointer to the parent SLA object
+ * * NULL - when parent cannot be found
+ */
+static struct rl_sla *find_parent(struct adf_rl *rl_data,
+ struct adf_rl_sla_input_data *sla_in)
+{
+ int input_parent_id = sla_in->parent_id;
+ struct rl_sla *root = NULL;
+ struct rl_sla *parent_sla;
+ int i;
+
+ if (sla_in->type == RL_ROOT)
+ return NULL;
+
+ if (input_parent_id > RL_PARENT_DEFAULT_ID) {
+ parent_sla = rl_data->sla[input_parent_id];
+ /*
+ * SLA can be a parent if it has the same service as the child
+ * and its type is higher in the hierarchy,
+ * for example the parent type of a LEAF must be a CLUSTER.
+ */
+ if (parent_sla && parent_sla->srv == sla_in->srv &&
+ parent_sla->type == sla_in->type - 1)
+ return parent_sla;
+
+ return NULL;
+ }
+
+ /* If input_parent_id is not valid, get root for this service type. */
+ for (i = 0; i < RL_ROOT_MAX; i++) {
+ if (rl_data->root[i] && rl_data->root[i]->srv == sla_in->srv) {
+ root = rl_data->root[i];
+ break;
+ }
+ }
+
+ if (!root)
+ return NULL;
+
+ /*
+ * If the type of this SLA is cluster, then return the root.
+ * Otherwise, find the default (i.e. first) cluster for this service.
+ */
+ if (sla_in->type == RL_CLUSTER)
+ return root;
+
+ for (i = 0; i < RL_CLUSTER_MAX; i++) {
+ if (rl_data->cluster[i] && rl_data->cluster[i]->parent == root)
+ return rl_data->cluster[i];
+ }
+
+ return NULL;
+}
+
+static enum adf_cfg_service_type srv_to_cfg_svc_type(enum adf_base_services rl_srv)
+{
+ switch (rl_srv) {
+ case ADF_SVC_ASYM:
+ return ASYM;
+ case ADF_SVC_SYM:
+ return SYM;
+ case ADF_SVC_DC:
+ return COMP;
+ default:
+ return UNUSED;
+ }
+}
+
+/**
+ * get_sla_arr_of_type() - Returns a pointer to SLA type specific array
+ * @rl_data: pointer to ratelimiting data
+ * @type: SLA type
+ * @sla_arr: pointer to variable where requested pointer will be stored
+ *
+ * Return: Max number of elements allowed for the returned array
+ */
+static u32 get_sla_arr_of_type(struct adf_rl *rl_data, enum rl_node_type type,
+ struct rl_sla ***sla_arr)
+{
+ switch (type) {
+ case RL_LEAF:
+ *sla_arr = rl_data->leaf;
+ return RL_LEAF_MAX;
+ case RL_CLUSTER:
+ *sla_arr = rl_data->cluster;
+ return RL_CLUSTER_MAX;
+ case RL_ROOT:
+ *sla_arr = rl_data->root;
+ return RL_ROOT_MAX;
+ default:
+ *sla_arr = NULL;
+ return 0;
+ }
+}
+
+static bool is_service_enabled(struct adf_accel_dev *accel_dev,
+ enum adf_base_services rl_srv)
+{
+ enum adf_cfg_service_type arb_srv = srv_to_cfg_svc_type(rl_srv);
+ struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev);
+ u8 rps_per_bundle = hw_data->num_banks_per_vf;
+ int i;
+
+ for (i = 0; i < rps_per_bundle; i++) {
+ if (GET_SRV_TYPE(accel_dev, i) == arb_srv)
+ return true;
+ }
+
+ return false;
+}
+
+/**
+ * prepare_rp_ids() - Creates an array of ring pair IDs from bitmask
+ * @accel_dev: pointer to acceleration device structure
+ * @sla: SLA object data where result will be written
+ * @rp_mask: bitmask of ring pair IDs
+ *
+ * Function tries to convert provided bitmap to an array of IDs. It checks if
+ * RPs aren't in use, are assigned to SLA service or if a number of provided
+ * IDs is not too big. If successful, writes the result into the field
+ * sla->ring_pairs_cnt.
+ *
+ * Return:
+ * * 0 - ok
+ * * -EINVAL - ring pairs array cannot be created from provided mask
+ */
+static int prepare_rp_ids(struct adf_accel_dev *accel_dev, struct rl_sla *sla,
+ const unsigned long rp_mask)
+{
+ enum adf_cfg_service_type arb_srv = srv_to_cfg_svc_type(sla->srv);
+ u16 rps_per_bundle = GET_HW_DATA(accel_dev)->num_banks_per_vf;
+ bool *rp_in_use = accel_dev->rate_limiting->rp_in_use;
+ size_t rp_cnt_max = ARRAY_SIZE(sla->ring_pairs_ids);
+ u16 rp_id_max = GET_HW_DATA(accel_dev)->num_banks;
+ u16 cnt = 0;
+ u16 rp_id;
+
+ for_each_set_bit(rp_id, &rp_mask, rp_id_max) {
+ if (cnt >= rp_cnt_max) {
+ dev_notice(&GET_DEV(accel_dev),
+ "Assigned more ring pairs than supported");
+ return -EINVAL;
+ }
+
+ if (rp_in_use[rp_id]) {
+ dev_notice(&GET_DEV(accel_dev),
+ "RP %u already assigned to other SLA", rp_id);
+ return -EINVAL;
+ }
+
+ if (GET_SRV_TYPE(accel_dev, rp_id % rps_per_bundle) != arb_srv) {
+ dev_notice(&GET_DEV(accel_dev),
+ "RP %u does not support SLA service", rp_id);
+ return -EINVAL;
+ }
+
+ sla->ring_pairs_ids[cnt++] = rp_id;
+ }
+
+ sla->ring_pairs_cnt = cnt;
+
+ return 0;
+}
+
+static void mark_rps_usage(struct rl_sla *sla, bool *rp_in_use, bool used)
+{
+ u16 rp_id;
+ int i;
+
+ for (i = 0; i < sla->ring_pairs_cnt; i++) {
+ rp_id = sla->ring_pairs_ids[i];
+ rp_in_use[rp_id] = used;
+ }
+}
+
+static void assign_rps_to_leaf(struct adf_accel_dev *accel_dev,
+ struct rl_sla *sla, bool clear)
+{
+ struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev);
+ void __iomem *pmisc_addr = adf_get_pmisc_base(accel_dev);
+ u32 base_offset = hw_data->rl_data.r2l_offset;
+ u32 node_id = clear ? 0U : (sla->node_id & LEAF_MASK);
+ u32 offset;
+ int i;
+
+ for (i = 0; i < sla->ring_pairs_cnt; i++) {
+ offset = base_offset + (RL_CSR_SIZE * sla->ring_pairs_ids[i]);
+ ADF_CSR_WR(pmisc_addr, offset, node_id);
+ }
+}
+
+static void assign_leaf_to_cluster(struct adf_accel_dev *accel_dev,
+ struct rl_sla *sla, bool clear)
+{
+ struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev);
+ void __iomem *pmisc_addr = adf_get_pmisc_base(accel_dev);
+ u32 base_offset = hw_data->rl_data.l2c_offset;
+ u32 node_id = sla->node_id & LEAF_MASK;
+ u32 parent_id = clear ? 0U : (sla->parent->node_id & CLUSTER_MASK);
+ u32 offset;
+
+ offset = base_offset + (RL_CSR_SIZE * node_id);
+ ADF_CSR_WR(pmisc_addr, offset, parent_id);
+}
+
+static void assign_cluster_to_root(struct adf_accel_dev *accel_dev,
+ struct rl_sla *sla, bool clear)
+{
+ struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev);
+ void __iomem *pmisc_addr = adf_get_pmisc_base(accel_dev);
+ u32 base_offset = hw_data->rl_data.c2s_offset;
+ u32 node_id = sla->node_id & CLUSTER_MASK;
+ u32 parent_id = clear ? 0U : (sla->parent->node_id & ROOT_MASK);
+ u32 offset;
+
+ offset = base_offset + (RL_CSR_SIZE * node_id);
+ ADF_CSR_WR(pmisc_addr, offset, parent_id);
+}
+
+static void assign_node_to_parent(struct adf_accel_dev *accel_dev,
+ struct rl_sla *sla, bool clear_assignment)
+{
+ switch (sla->type) {
+ case RL_LEAF:
+ assign_rps_to_leaf(accel_dev, sla, clear_assignment);
+ assign_leaf_to_cluster(accel_dev, sla, clear_assignment);
+ break;
+ case RL_CLUSTER:
+ assign_cluster_to_root(accel_dev, sla, clear_assignment);
+ break;
+ default:
+ break;
+ }
+}
+
+/**
+ * can_parent_afford_sla() - Verifies if parent allows to create an SLA
+ * @sla_in: pointer to user input data for a new SLA
+ * @sla_parent: pointer to parent SLA object
+ * @sla_cir: current child CIR value (only for update)
+ * @is_update: request is a update
+ *
+ * Algorithm verifies if parent has enough remaining budget to take assignment
+ * of a child with provided parameters. In update case current CIR value must be
+ * returned to budget first.
+ * PIR value cannot exceed the PIR assigned to parent.
+ *
+ * Return:
+ * * true - SLA can be created
+ * * false - SLA cannot be created
+ */
+static bool can_parent_afford_sla(struct adf_rl_sla_input_data *sla_in,
+ struct rl_sla *sla_parent, u32 sla_cir,
+ bool is_update)
+{
+ u32 rem_cir = sla_parent->rem_cir;
+
+ if (is_update)
+ rem_cir += sla_cir;
+
+ if (sla_in->cir > rem_cir || sla_in->pir > sla_parent->pir)
+ return false;
+
+ return true;
+}
+
+/**
+ * can_node_afford_update() - Verifies if SLA can be updated with input data
+ * @sla_in: pointer to user input data for a new SLA
+ * @sla: pointer to SLA object selected for update
+ *
+ * Algorithm verifies if a new CIR value is big enough to satisfy currently
+ * assigned child SLAs and if PIR can be updated
+ *
+ * Return:
+ * * true - SLA can be updated
+ * * false - SLA cannot be updated
+ */
+static bool can_node_afford_update(struct adf_rl_sla_input_data *sla_in,
+ struct rl_sla *sla)
+{
+ u32 cir_in_use = sla->cir - sla->rem_cir;
+
+ /* new CIR cannot be smaller then currently consumed value */
+ if (cir_in_use > sla_in->cir)
+ return false;
+
+ /* PIR of root/cluster cannot be reduced in node with assigned children */
+ if (sla_in->pir < sla->pir && sla->type != RL_LEAF && cir_in_use > 0)
+ return false;
+
+ return true;
+}
+
+static bool is_enough_budget(struct adf_rl *rl_data, struct rl_sla *sla,
+ struct adf_rl_sla_input_data *sla_in,
+ bool is_update)
+{
+ u32 max_val = rl_data->device_data->scale_ref;
+ struct rl_sla *parent = sla->parent;
+ bool ret = true;
+
+ if (sla_in->cir > max_val || sla_in->pir > max_val)
+ ret = false;
+
+ switch (sla->type) {
+ case RL_LEAF:
+ ret &= can_parent_afford_sla(sla_in, parent, sla->cir,
+ is_update);
+ break;
+ case RL_CLUSTER:
+ ret &= can_parent_afford_sla(sla_in, parent, sla->cir,
+ is_update);
+
+ if (is_update)
+ ret &= can_node_afford_update(sla_in, sla);
+
+ break;
+ case RL_ROOT:
+ if (is_update)
+ ret &= can_node_afford_update(sla_in, sla);
+
+ break;
+ default:
+ ret = false;
+ break;
+ }
+
+ return ret;
+}
+
+static void update_budget(struct rl_sla *sla, u32 old_cir, bool is_update)
+{
+ u32 new_rem;
+
+ switch (sla->type) {
+ case RL_LEAF:
+ if (is_update)
+ sla->parent->rem_cir += old_cir;
+
+ sla->parent->rem_cir -= sla->cir;
+ sla->rem_cir = 0;
+ break;
+ case RL_CLUSTER:
+ if (is_update) {
+ sla->parent->rem_cir += old_cir;
+ new_rem = sla->cir - (old_cir - sla->rem_cir);
+ sla->rem_cir = new_rem;
+ } else {
+ sla->rem_cir = sla->cir;
+ }
+
+ sla->parent->rem_cir -= sla->cir;
+ break;
+ case RL_ROOT:
+ if (is_update) {
+ new_rem = sla->cir - (old_cir - sla->rem_cir);
+ sla->rem_cir = new_rem;
+ } else {
+ sla->rem_cir = sla->cir;
+ }
+ break;
+ default:
+ break;
+ }
+}
+
+/**
+ * get_next_free_sla_id() - finds next free ID in the SLA array
+ * @rl_data: Pointer to ratelimiting data structure
+ *
+ * Return:
+ * * 0 : RL_NODES_CNT_MAX - correct ID
+ * * -ENOSPC - all SLA slots are in use
+ */
+static int get_next_free_sla_id(struct adf_rl *rl_data)
+{
+ int i = 0;
+
+ while (i < RL_NODES_CNT_MAX && rl_data->sla[i++])
+ ;
+
+ if (i == RL_NODES_CNT_MAX)
+ return -ENOSPC;
+
+ return i - 1;
+}
+
+/**
+ * get_next_free_node_id() - finds next free ID in the array of that node type
+ * @rl_data: Pointer to ratelimiting data structure
+ * @sla: Pointer to SLA object for which the ID is searched
+ *
+ * Return:
+ * * 0 : RL_[NODE_TYPE]_MAX - correct ID
+ * * -ENOSPC - all slots of that type are in use
+ */
+static int get_next_free_node_id(struct adf_rl *rl_data, struct rl_sla *sla)
+{
+ struct adf_hw_device_data *hw_device = GET_HW_DATA(rl_data->accel_dev);
+ int max_id, i, step, rp_per_leaf;
+ struct rl_sla **sla_list;
+
+ rp_per_leaf = hw_device->num_banks / hw_device->num_banks_per_vf;
+
+ /*
+ * Static nodes mapping:
+ * root0 - cluster[0,4,8,12] - leaf[0-15]
+ * root1 - cluster[1,5,9,13] - leaf[16-31]
+ * root2 - cluster[2,6,10,14] - leaf[32-47]
+ */
+ switch (sla->type) {
+ case RL_LEAF:
+ i = sla->srv * rp_per_leaf;
+ step = 1;
+ max_id = i + rp_per_leaf;
+ sla_list = rl_data->leaf;
+ break;
+ case RL_CLUSTER:
+ i = sla->srv;
+ step = 4;
+ max_id = RL_CLUSTER_MAX;
+ sla_list = rl_data->cluster;
+ break;
+ case RL_ROOT:
+ return sla->srv;
+ default:
+ return -EINVAL;
+ }
+
+ while (i < max_id && sla_list[i])
+ i += step;
+
+ if (i >= max_id)
+ return -ENOSPC;
+
+ return i;
+}
+
+u32 adf_rl_calculate_slice_tokens(struct adf_accel_dev *accel_dev, u32 sla_val,
+ enum adf_base_services svc_type)
+{
+ struct adf_rl_hw_data *device_data = &accel_dev->hw_device->rl_data;
+ struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev);
+ u64 avail_slice_cycles, allocated_tokens;
+
+ if (!sla_val)
+ return 0;
+
+ avail_slice_cycles = hw_data->clock_frequency;
+
+ switch (svc_type) {
+ case ADF_SVC_ASYM:
+ avail_slice_cycles *= device_data->slices.pke_cnt;
+ break;
+ case ADF_SVC_SYM:
+ avail_slice_cycles *= device_data->slices.cph_cnt;
+ break;
+ case ADF_SVC_DC:
+ avail_slice_cycles *= device_data->slices.dcpr_cnt;
+ break;
+ default:
+ break;
+ }
+
+ do_div(avail_slice_cycles, device_data->scan_interval);
+ allocated_tokens = avail_slice_cycles * sla_val;
+ do_div(allocated_tokens, device_data->scale_ref);
+
+ return allocated_tokens;
+}
+
+u32 adf_rl_calculate_ae_cycles(struct adf_accel_dev *accel_dev, u32 sla_val,
+ enum adf_base_services svc_type)
+{
+ struct adf_rl_hw_data *device_data = &accel_dev->hw_device->rl_data;
+ struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev);
+ u64 allocated_ae_cycles, avail_ae_cycles;
+
+ if (!sla_val)
+ return 0;
+
+ avail_ae_cycles = hw_data->clock_frequency;
+ avail_ae_cycles *= hw_data->get_num_aes(hw_data) - 1;
+ do_div(avail_ae_cycles, device_data->scan_interval);
+
+ sla_val *= device_data->max_tp[svc_type];
+ sla_val /= device_data->scale_ref;
+
+ allocated_ae_cycles = (sla_val * avail_ae_cycles);
+ do_div(allocated_ae_cycles, device_data->max_tp[svc_type]);
+
+ return allocated_ae_cycles;
+}
+
+u32 adf_rl_calculate_pci_bw(struct adf_accel_dev *accel_dev, u32 sla_val,
+ enum adf_base_services svc_type, bool is_bw_out)
+{
+ struct adf_rl_hw_data *device_data = &accel_dev->hw_device->rl_data;
+ u64 sla_to_bytes, allocated_bw, sla_scaled;
+
+ if (!sla_val)
+ return 0;
+
+ sla_to_bytes = sla_val;
+ sla_to_bytes *= device_data->max_tp[svc_type];
+ do_div(sla_to_bytes, device_data->scale_ref);
+
+ sla_to_bytes *= (svc_type == ADF_SVC_ASYM) ? RL_TOKEN_ASYM_SIZE :
+ BYTES_PER_MBIT;
+ if (svc_type == ADF_SVC_DC && is_bw_out)
+ sla_to_bytes *= device_data->slices.dcpr_cnt -
+ device_data->dcpr_correction;
+
+ sla_scaled = sla_to_bytes * device_data->pcie_scale_mul;
+ do_div(sla_scaled, device_data->pcie_scale_div);
+ allocated_bw = sla_scaled;
+ do_div(allocated_bw, RL_TOKEN_PCIE_SIZE);
+ do_div(allocated_bw, device_data->scan_interval);
+
+ return allocated_bw;
+}
+
+/**
+ * add_new_sla_entry() - creates a new SLA object and fills it with user data
+ * @accel_dev: pointer to acceleration device structure
+ * @sla_in: pointer to user input data for a new SLA
+ * @sla_out: Pointer to variable that will contain the address of a new
+ * SLA object if the operation succeeds
+ *
+ * Return:
+ * * 0 - ok
+ * * -ENOMEM - memory allocation failed
+ * * -EINVAL - invalid user input
+ * * -ENOSPC - all available SLAs are in use
+ */
+static int add_new_sla_entry(struct adf_accel_dev *accel_dev,
+ struct adf_rl_sla_input_data *sla_in,
+ struct rl_sla **sla_out)
+{
+ struct adf_rl *rl_data = accel_dev->rate_limiting;
+ struct rl_sla *sla;
+ int ret = 0;
+
+ sla = kzalloc(sizeof(*sla), GFP_KERNEL);
+ if (!sla) {
+ ret = -ENOMEM;
+ goto ret_err;
+ }
+ *sla_out = sla;
+
+ if (!is_service_enabled(accel_dev, sla_in->srv)) {
+ dev_notice(&GET_DEV(accel_dev),
+ "Provided service is not enabled\n");
+ ret = -EINVAL;
+ goto ret_err;
+ }
+
+ sla->srv = sla_in->srv;
+ sla->type = sla_in->type;
+ ret = get_next_free_node_id(rl_data, sla);
+ if (ret < 0) {
+ dev_notice(&GET_DEV(accel_dev),
+ "Exceeded number of available nodes for that service\n");
+ goto ret_err;
+ }
+ sla->node_id = ret;
+
+ ret = get_next_free_sla_id(rl_data);
+ if (ret < 0) {
+ dev_notice(&GET_DEV(accel_dev),
+ "Allocated maximum SLAs number\n");
+ goto ret_err;
+ }
+ sla->sla_id = ret;
+
+ sla->parent = find_parent(rl_data, sla_in);
+ if (!sla->parent && sla->type != RL_ROOT) {
+ if (sla_in->parent_id != RL_PARENT_DEFAULT_ID)
+ dev_notice(&GET_DEV(accel_dev),
+ "Provided parent ID does not exist or cannot be parent for this SLA.");
+ else
+ dev_notice(&GET_DEV(accel_dev),
+ "Unable to find parent node for this service. Is service enabled?");
+ ret = -EINVAL;
+ goto ret_err;
+ }
+
+ if (sla->type == RL_LEAF) {
+ ret = prepare_rp_ids(accel_dev, sla, sla_in->rp_mask);
+ if (!sla->ring_pairs_cnt || ret) {
+ dev_notice(&GET_DEV(accel_dev),
+ "Unable to find ring pairs to assign to the leaf");
+ if (!ret)
+ ret = -EINVAL;
+
+ goto ret_err;
+ }
+ }
+
+ ret = 0;
+
+ret_err:
+ /* Allocated sla will be freed at the bottom of calling function */
+ return ret;
+}
+
+static int initialize_default_nodes(struct adf_accel_dev *accel_dev)
+{
+ struct adf_rl *rl_data = accel_dev->rate_limiting;
+ struct adf_rl_hw_data *device_data = rl_data->device_data;
+ struct adf_rl_sla_input_data sla_in = { };
+ int ret = 0;
+ int i;
+
+ /* Init root for each enabled service */
+ sla_in.type = RL_ROOT;
+ sla_in.parent_id = RL_PARENT_DEFAULT_ID;
+
+ for (i = 0; i < ADF_SVC_NONE; i++) {
+ if (!is_service_enabled(accel_dev, i))
+ continue;
+
+ sla_in.cir = device_data->scale_ref;
+ sla_in.pir = sla_in.cir;
+ sla_in.srv = i;
+
+ ret = adf_rl_add_sla(accel_dev, &sla_in);
+ if (ret)
+ goto err_ret;
+ }
+
+ /* Init default cluster for each root */
+ sla_in.type = RL_CLUSTER;
+ for (i = 0; i < ADF_SVC_NONE; i++) {
+ if (!rl_data->root[i])
+ continue;
+
+ sla_in.cir = rl_data->root[i]->cir;
+ sla_in.pir = sla_in.cir;
+ sla_in.srv = rl_data->root[i]->srv;
+
+ ret = adf_rl_add_sla(accel_dev, &sla_in);
+ if (ret)
+ goto err_ret;
+ }
+
+ return 0;
+
+err_ret:
+ dev_notice(&GET_DEV(accel_dev),
+ "Initialization of default nodes failed\n");
+ return ret;
+}
+
+static void clear_sla(struct adf_rl *rl_data, struct rl_sla *sla)
+{
+ bool *rp_in_use = rl_data->rp_in_use;
+ struct rl_sla **sla_type_arr = NULL;
+ int i, sla_id, node_id;
+ u32 old_cir;
+
+ sla_id = sla->sla_id;
+ node_id = sla->node_id;
+ old_cir = sla->cir;
+ sla->cir = 0;
+ sla->pir = 0;
+
+ for (i = 0; i < sla->ring_pairs_cnt; i++)
+ rp_in_use[sla->ring_pairs_ids[i]] = false;
+
+ update_budget(sla, old_cir, true);
+ get_sla_arr_of_type(rl_data, sla->type, &sla_type_arr);
+ assign_node_to_parent(rl_data->accel_dev, sla, true);
+ adf_rl_send_admin_delete_msg(rl_data->accel_dev, node_id, sla->type);
+ mark_rps_usage(sla, rl_data->rp_in_use, false);
+
+ kfree(sla);
+ rl_data->sla[sla_id] = NULL;
+ sla_type_arr[node_id] = NULL;
+}
+
+/**
+ * add_update_sla() - handles the creation and the update of an SLA
+ * @accel_dev: pointer to acceleration device structure
+ * @sla_in: pointer to user input data for a new/updated SLA
+ * @is_update: flag to indicate if this is an update or an add operation
+ *
+ * Return:
+ * * 0 - ok
+ * * -ENOMEM - memory allocation failed
+ * * -EINVAL - user input data cannot be used to create SLA
+ * * -ENOSPC - all available SLAs are in use
+ */
+static int add_update_sla(struct adf_accel_dev *accel_dev,
+ struct adf_rl_sla_input_data *sla_in, bool is_update)
+{
+ struct adf_rl *rl_data = accel_dev->rate_limiting;
+ struct rl_sla **sla_type_arr = NULL;
+ struct rl_sla *sla = NULL;
+ u32 old_cir = 0;
+ int ret;
+
+ if (!sla_in) {
+ dev_warn(&GET_DEV(accel_dev),
+ "SLA input data pointer is missing\n");
+ ret = -EFAULT;
+ goto ret_err;
+ }
+
+ /* Input validation */
+ ret = validate_user_input(accel_dev, sla_in, is_update);
+ if (ret)
+ goto ret_err;
+
+ mutex_lock(&rl_data->rl_lock);
+
+ if (is_update) {
+ ret = validate_sla_id(accel_dev, sla_in->sla_id);
+ if (ret)
+ goto ret_err;
+
+ sla = rl_data->sla[sla_in->sla_id];
+ old_cir = sla->cir;
+ } else {
+ ret = add_new_sla_entry(accel_dev, sla_in, &sla);
+ if (ret)
+ goto ret_err;
+ }
+
+ if (!is_enough_budget(rl_data, sla, sla_in, is_update)) {
+ dev_notice(&GET_DEV(accel_dev),
+ "Input value exceeds the remaining budget%s\n",
+ is_update ? " or more budget is already in use" : "");
+ ret = -EINVAL;
+ goto ret_err;
+ }
+ sla->cir = sla_in->cir;
+ sla->pir = sla_in->pir;
+
+ /* Apply SLA */
+ assign_node_to_parent(accel_dev, sla, false);
+ ret = adf_rl_send_admin_add_update_msg(accel_dev, sla, is_update);
+ if (ret) {
+ dev_notice(&GET_DEV(accel_dev),
+ "Failed to apply an SLA\n");
+ goto ret_err;
+ }
+ update_budget(sla, old_cir, is_update);
+
+ if (!is_update) {
+ mark_rps_usage(sla, rl_data->rp_in_use, true);
+ get_sla_arr_of_type(rl_data, sla->type, &sla_type_arr);
+ sla_type_arr[sla->node_id] = sla;
+ rl_data->sla[sla->sla_id] = sla;
+ }
+
+ sla_in->sla_id = sla->sla_id;
+ goto ret_ok;
+
+ret_err:
+ if (!is_update && sla) {
+ sla_in->sla_id = -1;
+ kfree(sla);
+ }
+ret_ok:
+ mutex_unlock(&rl_data->rl_lock);
+ return ret;
+}
+
+/**
+ * adf_rl_add_sla() - handles the creation of an SLA
+ * @accel_dev: pointer to acceleration device structure
+ * @sla_in: pointer to user input data required to add an SLA
+ *
+ * Return:
+ * * 0 - ok
+ * * -ENOMEM - memory allocation failed
+ * * -EINVAL - invalid user input
+ * * -ENOSPC - all available SLAs are in use
+ */
+int adf_rl_add_sla(struct adf_accel_dev *accel_dev,
+ struct adf_rl_sla_input_data *sla_in)
+{
+ return add_update_sla(accel_dev, sla_in, false);
+}
+
+/**
+ * adf_rl_update_sla() - handles the update of an SLA
+ * @accel_dev: pointer to acceleration device structure
+ * @sla_in: pointer to user input data required to update an SLA
+ *
+ * Return:
+ * * 0 - ok
+ * * -EINVAL - user input data cannot be used to update SLA
+ */
+int adf_rl_update_sla(struct adf_accel_dev *accel_dev,
+ struct adf_rl_sla_input_data *sla_in)
+{
+ return add_update_sla(accel_dev, sla_in, true);
+}
+
+/**
+ * adf_rl_get_sla() - returns an existing SLA data
+ * @accel_dev: pointer to acceleration device structure
+ * @sla_in: pointer to user data where SLA info will be stored
+ *
+ * The sla_id for which data are requested should be set in sla_id structure
+ *
+ * Return:
+ * * 0 - ok
+ * * -EINVAL - provided sla_id does not exist
+ */
+int adf_rl_get_sla(struct adf_accel_dev *accel_dev,
+ struct adf_rl_sla_input_data *sla_in)
+{
+ struct rl_sla *sla;
+ int ret, i;
+
+ ret = validate_sla_id(accel_dev, sla_in->sla_id);
+ if (ret)
+ return ret;
+
+ sla = accel_dev->rate_limiting->sla[sla_in->sla_id];
+ sla_in->type = sla->type;
+ sla_in->srv = sla->srv;
+ sla_in->cir = sla->cir;
+ sla_in->pir = sla->pir;
+ sla_in->rp_mask = 0U;
+ if (sla->parent)
+ sla_in->parent_id = sla->parent->sla_id;
+ else
+ sla_in->parent_id = RL_PARENT_DEFAULT_ID;
+
+ for (i = 0; i < sla->ring_pairs_cnt; i++)
+ sla_in->rp_mask |= BIT(sla->ring_pairs_ids[i]);
+
+ return 0;
+}
+
+/**
+ * adf_rl_get_capability_remaining() - returns the remaining SLA value (CIR) for
+ * selected service or provided sla_id
+ * @accel_dev: pointer to acceleration device structure
+ * @srv: service ID for which capability is requested
+ * @sla_id: ID of the cluster or root to which we want assign a new SLA
+ *
+ * Check if the provided SLA id is valid. If it is and the service matches
+ * the requested service and the type is cluster or root, return the remaining
+ * capability.
+ * If the provided ID does not match the service or type, return the remaining
+ * capacity of the default cluster for that service.
+ *
+ * Return:
+ * * Positive value - correct remaining value
+ * * -EINVAL - algorithm cannot find a remaining value for provided data
+ */
+int adf_rl_get_capability_remaining(struct adf_accel_dev *accel_dev,
+ enum adf_base_services srv, int sla_id)
+{
+ struct adf_rl *rl_data = accel_dev->rate_limiting;
+ struct rl_sla *sla = NULL;
+ int i;
+
+ if (srv >= ADF_SVC_NONE)
+ return -EINVAL;
+
+ if (sla_id > RL_SLA_EMPTY_ID && !validate_sla_id(accel_dev, sla_id)) {
+ sla = rl_data->sla[sla_id];
+
+ if (sla->srv == srv && sla->type <= RL_CLUSTER)
+ goto ret_ok;
+ }
+
+ for (i = 0; i < RL_CLUSTER_MAX; i++) {
+ if (!rl_data->cluster[i])
+ continue;
+
+ if (rl_data->cluster[i]->srv == srv) {
+ sla = rl_data->cluster[i];
+ goto ret_ok;
+ }
+ }
+
+ return -EINVAL;
+ret_ok:
+ return sla->rem_cir;
+}
+
+/**
+ * adf_rl_remove_sla() - removes provided sla_id
+ * @accel_dev: pointer to acceleration device structure
+ * @sla_id: ID of the cluster or root to which we want assign an new SLA
+ *
+ * Return:
+ * * 0 - ok
+ * * -EINVAL - wrong sla_id or it still have assigned children
+ */
+int adf_rl_remove_sla(struct adf_accel_dev *accel_dev, u32 sla_id)
+{
+ struct adf_rl *rl_data = accel_dev->rate_limiting;
+ struct rl_sla *sla;
+ int ret;
+
+ ret = validate_sla_id(accel_dev, sla_id);
+ if (ret)
+ return ret;
+
+ sla = rl_data->sla[sla_id];
+
+ if (sla->type < RL_LEAF && sla->rem_cir != sla->cir) {
+ dev_notice(&GET_DEV(accel_dev),
+ "To remove parent SLA all its children must be removed first");
+ return -EINVAL;
+ }
+
+ mutex_lock(&rl_data->rl_lock);
+ clear_sla(rl_data, sla);
+ mutex_unlock(&rl_data->rl_lock);
+
+ return 0;
+}
+
+/**
+ * adf_rl_remove_sla_all() - removes all SLAs from device
+ * @accel_dev: pointer to acceleration device structure
+ * @incl_default: set to true if default SLAs also should be removed
+ */
+void adf_rl_remove_sla_all(struct adf_accel_dev *accel_dev, bool incl_default)
+{
+ struct adf_rl *rl_data = accel_dev->rate_limiting;
+ int end_type = incl_default ? RL_ROOT : RL_LEAF;
+ struct rl_sla **sla_type_arr = NULL;
+ u32 max_id;
+ int i, j;
+
+ mutex_lock(&rl_data->rl_lock);
+
+ /* Unregister and remove all SLAs */
+ for (j = RL_LEAF; j >= end_type; j--) {
+ max_id = get_sla_arr_of_type(rl_data, j, &sla_type_arr);
+
+ for (i = 0; i < max_id; i++) {
+ if (!sla_type_arr[i])
+ continue;
+
+ clear_sla(rl_data, sla_type_arr[i]);
+ }
+ }
+
+ mutex_unlock(&rl_data->rl_lock);
+}
+
+int adf_rl_init(struct adf_accel_dev *accel_dev)
+{
+ struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev);
+ struct adf_rl_hw_data *rl_hw_data = &hw_data->rl_data;
+ struct adf_rl *rl;
+ int ret = 0;
+
+ /* Validate device parameters */
+ if (RL_VALIDATE_NON_ZERO(rl_hw_data->max_tp[ADF_SVC_ASYM]) ||
+ RL_VALIDATE_NON_ZERO(rl_hw_data->max_tp[ADF_SVC_SYM]) ||
+ RL_VALIDATE_NON_ZERO(rl_hw_data->max_tp[ADF_SVC_DC]) ||
+ RL_VALIDATE_NON_ZERO(rl_hw_data->scan_interval) ||
+ RL_VALIDATE_NON_ZERO(rl_hw_data->pcie_scale_div) ||
+ RL_VALIDATE_NON_ZERO(rl_hw_data->pcie_scale_mul) ||
+ RL_VALIDATE_NON_ZERO(rl_hw_data->scale_ref)) {
+ ret = -EOPNOTSUPP;
+ goto err_ret;
+ }
+
+ rl = kzalloc(sizeof(*rl), GFP_KERNEL);
+ if (!rl) {
+ ret = -ENOMEM;
+ goto err_ret;
+ }
+
+ mutex_init(&rl->rl_lock);
+ rl->device_data = &accel_dev->hw_device->rl_data;
+ rl->accel_dev = accel_dev;
+ accel_dev->rate_limiting = rl;
+
+err_ret:
+ return ret;
+}
+
+int adf_rl_start(struct adf_accel_dev *accel_dev)
+{
+ struct adf_rl_hw_data *rl_hw_data = &GET_HW_DATA(accel_dev)->rl_data;
+ void __iomem *pmisc_addr = adf_get_pmisc_base(accel_dev);
+ u16 fw_caps = GET_HW_DATA(accel_dev)->fw_capabilities;
+ int ret;
+
+ if (!accel_dev->rate_limiting) {
+ ret = -EOPNOTSUPP;
+ goto ret_err;
+ }
+
+ if ((fw_caps & RL_CAPABILITY_MASK) != RL_CAPABILITY_VALUE) {
+ dev_info(&GET_DEV(accel_dev), "not supported\n");
+ ret = -EOPNOTSUPP;
+ goto ret_free;
+ }
+
+ ADF_CSR_WR(pmisc_addr, rl_hw_data->pciin_tb_offset,
+ RL_TOKEN_GRANULARITY_PCIEIN_BUCKET);
+ ADF_CSR_WR(pmisc_addr, rl_hw_data->pciout_tb_offset,
+ RL_TOKEN_GRANULARITY_PCIEOUT_BUCKET);
+
+ ret = adf_rl_send_admin_init_msg(accel_dev, &rl_hw_data->slices);
+ if (ret) {
+ dev_err(&GET_DEV(accel_dev), "initialization failed\n");
+ goto ret_free;
+ }
+
+ ret = initialize_default_nodes(accel_dev);
+ if (ret) {
+ dev_err(&GET_DEV(accel_dev),
+ "failed to initialize default SLAs\n");
+ goto ret_sla_rm;
+ }
+
+ return 0;
+
+ret_sla_rm:
+ adf_rl_remove_sla_all(accel_dev, true);
+ret_free:
+ kfree(accel_dev->rate_limiting);
+ accel_dev->rate_limiting = NULL;
+ret_err:
+ return ret;
+}
+
+void adf_rl_stop(struct adf_accel_dev *accel_dev)
+{
+ if (!accel_dev->rate_limiting)
+ return;
+
+ adf_rl_remove_sla_all(accel_dev, true);
+}
+
+void adf_rl_exit(struct adf_accel_dev *accel_dev)
+{
+ if (!accel_dev->rate_limiting)
+ return;
+
+ kfree(accel_dev->rate_limiting);
+ accel_dev->rate_limiting = NULL;
+}
diff --git a/drivers/crypto/intel/qat/qat_common/adf_rl.h b/drivers/crypto/intel/qat/qat_common/adf_rl.h
new file mode 100644
index 000000000000..1ccb6613c92e
--- /dev/null
+++ b/drivers/crypto/intel/qat/qat_common/adf_rl.h
@@ -0,0 +1,169 @@
+/* SPDX-License-Identifier: GPL-2.0-only */
+/* Copyright(c) 2023 Intel Corporation */
+
+#ifndef ADF_RL_H_
+#define ADF_RL_H_
+
+#include <linux/mutex.h>
+#include <linux/types.h>
+
+struct adf_accel_dev;
+
+#define RL_ROOT_MAX 4
+#define RL_CLUSTER_MAX 16
+#define RL_LEAF_MAX 64
+#define RL_NODES_CNT_MAX (RL_ROOT_MAX + RL_CLUSTER_MAX + RL_LEAF_MAX)
+#define RL_RP_CNT_PER_LEAF_MAX 4U
+#define RL_RP_CNT_MAX 64
+#define RL_SLA_EMPTY_ID -1
+#define RL_PARENT_DEFAULT_ID -1
+
+enum rl_node_type {
+ RL_ROOT,
+ RL_CLUSTER,
+ RL_LEAF,
+};
+
+enum adf_base_services {
+ ADF_SVC_ASYM = 0,
+ ADF_SVC_SYM,
+ ADF_SVC_DC,
+ ADF_SVC_NONE,
+};
+
+/**
+ * struct adf_rl_sla_input_data - ratelimiting user input data structure
+ * @rp_mask: 64 bit bitmask of ring pair IDs which will be assigned to SLA.
+ * Eg. 0x5 -> RP0 and RP2 assigned; 0xA005 -> RP0,2,13,15 assigned.
+ * @sla_id: ID of current SLA for operations update, rm, get. For the add
+ * operation, this field will be updated with the ID of the newly
+ * added SLA
+ * @parent_id: ID of the SLA to which the current one should be assigned.
+ * Set to -1 to refer to the default parent.
+ * @cir: Committed information rate. Rate guaranteed to be achieved. Input value
+ * is expressed in permille scale, i.e. 1000 refers to the maximum
+ * device throughput for a selected service.
+ * @pir: Peak information rate. Maximum rate available that the SLA can achieve.
+ * Input value is expressed in permille scale, i.e. 1000 refers to
+ * the maximum device throughput for a selected service.
+ * @type: SLA type: root, cluster, node
+ * @srv: Service associated to the SLA: asym, sym dc.
+ *
+ * This structure is used to perform operations on an SLA.
+ * Depending on the operation, some of the parameters are ignored.
+ * The following list reports which parameters should be set for each operation.
+ * - add: all except sla_id
+ * - update: cir, pir, sla_id
+ * - rm: sla_id
+ * - rm_all: -
+ * - get: sla_id
+ * - get_capability_rem: srv, sla_id
+ */
+struct adf_rl_sla_input_data {
+ u64 rp_mask;
+ int sla_id;
+ int parent_id;
+ unsigned int cir;
+ unsigned int pir;
+ enum rl_node_type type;
+ enum adf_base_services srv;
+};
+
+struct rl_slice_cnt {
+ u8 dcpr_cnt;
+ u8 pke_cnt;
+ u8 cph_cnt;
+};
+
+struct adf_rl_hw_data {
+ u32 scale_ref;
+ u32 scan_interval;
+ u32 r2l_offset;
+ u32 l2c_offset;
+ u32 c2s_offset;
+ u32 pciin_tb_offset;
+ u32 pciout_tb_offset;
+ u32 pcie_scale_mul;
+ u32 pcie_scale_div;
+ u32 dcpr_correction;
+ u32 max_tp[RL_ROOT_MAX];
+ struct rl_slice_cnt slices;
+};
+
+/**
+ * struct adf_rl - ratelimiting data structure
+ * @accel_dev: pointer to acceleration device data
+ * @device_data: pointer to rate limiting data specific to a device type (or revision)
+ * @sla: array of pointers to SLA objects
+ * @root: array of pointers to root type SLAs, element number reflects node_id
+ * @cluster: array of pointers to cluster type SLAs, element number reflects node_id
+ * @leaf: array of pointers to leaf type SLAs, element number reflects node_id
+ * @rp_in_use: array of ring pair IDs already used in one of SLAs
+ * @rl_lock: mutex object which is protecting data in this structure
+ * @input: structure which is used for holding the data received from user
+ */
+struct adf_rl {
+ struct adf_accel_dev *accel_dev;
+ struct adf_rl_hw_data *device_data;
+ /* mapping sla_id to SLA objects */
+ struct rl_sla *sla[RL_NODES_CNT_MAX];
+ struct rl_sla *root[RL_ROOT_MAX];
+ struct rl_sla *cluster[RL_CLUSTER_MAX];
+ struct rl_sla *leaf[RL_LEAF_MAX];
+ bool rp_in_use[RL_RP_CNT_MAX];
+ /* Mutex protecting writing to SLAs lists */
+ struct mutex rl_lock;
+};
+
+/**
+ * struct rl_sla - SLA object data structure
+ * @parent: pointer to the parent SLA (root/cluster)
+ * @type: SLA type
+ * @srv: service associated with this SLA
+ * @sla_id: ID of the SLA, used as element number in SLA array and as identifier
+ * shared with the user
+ * @node_id: ID of node, each of SLA type have a separate ID list
+ * @cir: committed information rate
+ * @pir: peak information rate (PIR >= CIR)
+ * @rem_cir: if this SLA is a parent then this field represents a remaining
+ * value to be used by child SLAs.
+ * @ring_pairs_ids: array with numeric ring pairs IDs assigned to this SLA
+ * @ring_pairs_cnt: number of assigned ring pairs listed in the array above
+ */
+struct rl_sla {
+ struct rl_sla *parent;
+ enum rl_node_type type;
+ enum adf_base_services srv;
+ u32 sla_id;
+ u32 node_id;
+ u32 cir;
+ u32 pir;
+ u32 rem_cir;
+ u16 ring_pairs_ids[RL_RP_CNT_PER_LEAF_MAX];
+ u16 ring_pairs_cnt;
+};
+
+int adf_rl_add_sla(struct adf_accel_dev *accel_dev,
+ struct adf_rl_sla_input_data *sla_in);
+int adf_rl_update_sla(struct adf_accel_dev *accel_dev,
+ struct adf_rl_sla_input_data *sla_in);
+int adf_rl_get_sla(struct adf_accel_dev *accel_dev,
+ struct adf_rl_sla_input_data *sla_in);
+int adf_rl_get_capability_remaining(struct adf_accel_dev *accel_dev,
+ enum adf_base_services srv, int sla_id);
+int adf_rl_remove_sla(struct adf_accel_dev *accel_dev, u32 sla_id);
+void adf_rl_remove_sla_all(struct adf_accel_dev *accel_dev, bool incl_default);
+
+int adf_rl_init(struct adf_accel_dev *accel_dev);
+int adf_rl_start(struct adf_accel_dev *accel_dev);
+void adf_rl_stop(struct adf_accel_dev *accel_dev);
+void adf_rl_exit(struct adf_accel_dev *accel_dev);
+
+u32 adf_rl_calculate_pci_bw(struct adf_accel_dev *accel_dev, u32 sla_val,
+ enum adf_base_services svc_type, bool is_bw_out);
+u32 adf_rl_calculate_ae_cycles(struct adf_accel_dev *accel_dev, u32 sla_val,
+ enum adf_base_services svc_type);
+u32 adf_rl_calculate_slice_tokens(struct adf_accel_dev *accel_dev, u32 sla_val,
+ enum adf_base_services svc_type);
+
+#endif /* ADF_RL_H_ */
diff --git a/drivers/crypto/intel/qat/qat_common/adf_rl_admin.c b/drivers/crypto/intel/qat/qat_common/adf_rl_admin.c
new file mode 100644
index 000000000000..820fcb2d42bf
--- /dev/null
+++ b/drivers/crypto/intel/qat/qat_common/adf_rl_admin.c
@@ -0,0 +1,98 @@
+// SPDX-License-Identifier: GPL-2.0-only
+/* Copyright(c) 2023 Intel Corporation */
+
+#include <linux/dma-mapping.h>
+#include <linux/pci.h>
+
+#include "adf_admin.h"
+#include "adf_accel_devices.h"
+#include "adf_rl_admin.h"
+
+static void
+prep_admin_req_msg(struct rl_sla *sla, dma_addr_t dma_addr,
+ struct icp_qat_fw_init_admin_sla_config_params *fw_params,
+ struct icp_qat_fw_init_admin_req *req, bool is_update)
+{
+ req->cmd_id = is_update ? ICP_QAT_FW_RL_UPDATE : ICP_QAT_FW_RL_ADD;
+ req->init_cfg_ptr = dma_addr;
+ req->init_cfg_sz = sizeof(*fw_params);
+ req->node_id = sla->node_id;
+ req->node_type = sla->type;
+ req->rp_count = sla->ring_pairs_cnt;
+ req->svc_type = sla->srv;
+}
+
+static void
+prep_admin_req_params(struct adf_accel_dev *accel_dev, struct rl_sla *sla,
+ struct icp_qat_fw_init_admin_sla_config_params *fw_params)
+{
+ fw_params->pcie_in_cir =
+ adf_rl_calculate_pci_bw(accel_dev, sla->cir, sla->srv, false);
+ fw_params->pcie_in_pir =
+ adf_rl_calculate_pci_bw(accel_dev, sla->pir, sla->srv, false);
+ fw_params->pcie_out_cir =
+ adf_rl_calculate_pci_bw(accel_dev, sla->cir, sla->srv, true);
+ fw_params->pcie_out_pir =
+ adf_rl_calculate_pci_bw(accel_dev, sla->pir, sla->srv, true);
+
+ fw_params->slice_util_cir =
+ adf_rl_calculate_slice_tokens(accel_dev, sla->cir, sla->srv);
+ fw_params->slice_util_pir =
+ adf_rl_calculate_slice_tokens(accel_dev, sla->pir, sla->srv);
+
+ fw_params->ae_util_cir =
+ adf_rl_calculate_ae_cycles(accel_dev, sla->cir, sla->srv);
+ fw_params->ae_util_pir =
+ adf_rl_calculate_ae_cycles(accel_dev, sla->pir, sla->srv);
+
+ memcpy(fw_params->rp_ids, sla->ring_pairs_ids,
+ sizeof(sla->ring_pairs_ids));
+}
+
+int adf_rl_send_admin_init_msg(struct adf_accel_dev *accel_dev,
+ struct rl_slice_cnt *slices_int)
+{
+ struct icp_qat_fw_init_admin_slice_cnt slices_resp = { };
+ int ret;
+
+ ret = adf_send_admin_rl_init(accel_dev, &slices_resp);
+ if (ret)
+ goto err_ret;
+
+ slices_int->dcpr_cnt = slices_resp.dcpr_cnt;
+ slices_int->pke_cnt = slices_resp.pke_cnt;
+ /* For symmetric crypto, slice tokens are relative to the UCS slice */
+ slices_int->cph_cnt = slices_resp.ucs_cnt;
+
+err_ret:
+ return ret;
+}
+
+int adf_rl_send_admin_add_update_msg(struct adf_accel_dev *accel_dev,
+ struct rl_sla *sla, bool is_update)
+{
+ struct icp_qat_fw_init_admin_sla_config_params *fw_params;
+ struct icp_qat_fw_init_admin_req req = { };
+ dma_addr_t dma_addr;
+ int ret;
+
+ fw_params = dma_alloc_coherent(&GET_DEV(accel_dev), sizeof(*fw_params),
+ &dma_addr, GFP_KERNEL);
+ if (!fw_params)
+ return -ENOMEM;
+
+ prep_admin_req_params(accel_dev, sla, fw_params);
+ prep_admin_req_msg(sla, dma_addr, fw_params, &req, is_update);
+ ret = adf_send_admin_rl_add_update(accel_dev, &req);
+
+ dma_free_coherent(&GET_DEV(accel_dev), sizeof(*fw_params), fw_params,
+ dma_addr);
+
+ return ret;
+}
+
+int adf_rl_send_admin_delete_msg(struct adf_accel_dev *accel_dev, u16 node_id,
+ u8 node_type)
+{
+ return adf_send_admin_rl_delete(accel_dev, node_id, node_type);
+}
diff --git a/drivers/crypto/intel/qat/qat_common/adf_rl_admin.h b/drivers/crypto/intel/qat/qat_common/adf_rl_admin.h
new file mode 100644
index 000000000000..dd5419b7e896
--- /dev/null
+++ b/drivers/crypto/intel/qat/qat_common/adf_rl_admin.h
@@ -0,0 +1,18 @@
+/* SPDX-License-Identifier: GPL-2.0-only */
+/* Copyright(c) 2023 Intel Corporation */
+
+#ifndef ADF_RL_ADMIN_H_
+#define ADF_RL_ADMIN_H_
+
+#include <linux/types.h>
+
+#include "adf_rl.h"
+
+int adf_rl_send_admin_init_msg(struct adf_accel_dev *accel_dev,
+ struct rl_slice_cnt *slices_int);
+int adf_rl_send_admin_add_update_msg(struct adf_accel_dev *accel_dev,
+ struct rl_sla *sla, bool is_update);
+int adf_rl_send_admin_delete_msg(struct adf_accel_dev *accel_dev, u16 node_id,
+ u8 node_type);
+
+#endif /* ADF_RL_ADMIN_H_ */
diff --git a/drivers/crypto/intel/qat/qat_common/icp_qat_fw_init_admin.h b/drivers/crypto/intel/qat/qat_common/icp_qat_fw_init_admin.h
index e4de9a30e0bd..cd418b51d9f3 100644
--- a/drivers/crypto/intel/qat/qat_common/icp_qat_fw_init_admin.h
+++ b/drivers/crypto/intel/qat/qat_common/icp_qat_fw_init_admin.h
@@ -5,6 +5,8 @@

#include "icp_qat_fw.h"

+#define RL_MAX_RP_IDS 16
+
enum icp_qat_fw_init_admin_cmd_id {
ICP_QAT_FW_INIT_AE = 0,
ICP_QAT_FW_TRNG_ENABLE = 1,
@@ -19,10 +21,14 @@ enum icp_qat_fw_init_admin_cmd_id {
ICP_QAT_FW_CRYPTO_CAPABILITY_GET = 10,
ICP_QAT_FW_DC_CHAIN_INIT = 11,
ICP_QAT_FW_HEARTBEAT_TIMER_SET = 13,
+ ICP_QAT_FW_RL_INIT = 15,
ICP_QAT_FW_TIMER_GET = 19,
ICP_QAT_FW_CNV_STATS_GET = 20,
ICP_QAT_FW_PM_STATE_CONFIG = 128,
ICP_QAT_FW_PM_INFO = 129,
+ ICP_QAT_FW_RL_ADD = 134,
+ ICP_QAT_FW_RL_UPDATE = 135,
+ ICP_QAT_FW_RL_REMOVE = 136,
};

enum icp_qat_fw_init_admin_resp_status {
@@ -30,6 +36,30 @@ enum icp_qat_fw_init_admin_resp_status {
ICP_QAT_FW_INIT_RESP_STATUS_FAIL
};

+struct icp_qat_fw_init_admin_slice_cnt {
+ __u8 cpr_cnt;
+ __u8 xlt_cnt;
+ __u8 dcpr_cnt;
+ __u8 pke_cnt;
+ __u8 wat_cnt;
+ __u8 wcp_cnt;
+ __u8 ucs_cnt;
+ __u8 cph_cnt;
+ __u8 ath_cnt;
+};
+
+struct icp_qat_fw_init_admin_sla_config_params {
+ __u32 pcie_in_cir;
+ __u32 pcie_in_pir;
+ __u32 pcie_out_cir;
+ __u32 pcie_out_pir;
+ __u32 slice_util_cir;
+ __u32 slice_util_pir;
+ __u32 ae_util_cir;
+ __u32 ae_util_pir;
+ __u16 rp_ids[RL_MAX_RP_IDS];
+};
+
struct icp_qat_fw_init_admin_req {
__u16 init_cfg_sz;
__u8 resrvd1;
@@ -49,6 +79,13 @@ struct icp_qat_fw_init_admin_req {
struct {
__u32 heartbeat_ticks;
};
+ struct {
+ __u16 node_id;
+ __u8 node_type;
+ __u8 svc_type;
+ __u8 resrvd5[3];
+ __u8 rp_count;
+ };
__u32 idle_filter;
};

@@ -110,6 +147,7 @@ struct icp_qat_fw_init_admin_resp {
__u32 unsuccessful_count;
__u64 resrvd8;
};
+ struct icp_qat_fw_init_admin_slice_cnt slices;
__u16 fw_capabilities;
};
} __packed;
--
2.41.0

2023-10-11 12:36:50

by Damian Muszynski

[permalink] [raw]
Subject: [PATCH 11/11] crypto: qat - add num_rps sysfs attribute

From: Ciunas Bennett <[email protected]>

Add the attribute `num_rps` to the `qat` attribute group. This returns
the number of ring pairs that a single device has. This allows to know
the maximum value that can be set to the attribute `rp2svc`.

Signed-off-by: Ciunas Bennett <[email protected]>
Reviewed-by: Giovanni Cabiddu <[email protected]>
Reviewed-by: Damian Muszynski <[email protected]>
---
Documentation/ABI/testing/sysfs-driver-qat | 14 ++++++++++++++
drivers/crypto/intel/qat/qat_common/adf_sysfs.c | 14 ++++++++++++++
2 files changed, 28 insertions(+)

diff --git a/Documentation/ABI/testing/sysfs-driver-qat b/Documentation/ABI/testing/sysfs-driver-qat
index f24a5ddca94b..bbf329cf0d67 100644
--- a/Documentation/ABI/testing/sysfs-driver-qat
+++ b/Documentation/ABI/testing/sysfs-driver-qat
@@ -127,3 +127,17 @@ Description:
sym

This attribute is only available for qat_4xxx devices.
+
+What: /sys/bus/pci/devices/<BDF>/qat/num_rps
+Date: January 2024
+KernelVersion: 6.7
+Contact: [email protected]
+Description:
+ (RO) Returns the number of ring pairs that a single device has.
+
+ Example usage::
+
+ # cat /sys/bus/pci/devices/<BDF>/qat/num_rps
+ 64
+
+ This attribute is only available for qat_4xxx devices.
diff --git a/drivers/crypto/intel/qat/qat_common/adf_sysfs.c b/drivers/crypto/intel/qat/qat_common/adf_sysfs.c
index 9317127128a9..ddffc98119c6 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_sysfs.c
+++ b/drivers/crypto/intel/qat/qat_common/adf_sysfs.c
@@ -268,11 +268,25 @@ static ssize_t rp2srv_store(struct device *dev, struct device_attribute *attr,
}
static DEVICE_ATTR_RW(rp2srv);

+static ssize_t num_rps_show(struct device *dev, struct device_attribute *attr,
+ char *buf)
+{
+ struct adf_accel_dev *accel_dev;
+
+ accel_dev = adf_devmgr_pci_to_accel_dev(to_pci_dev(dev));
+ if (!accel_dev)
+ return -EINVAL;
+
+ return sysfs_emit(buf, "%u\n", GET_MAX_BANKS(accel_dev));
+}
+static DEVICE_ATTR_RO(num_rps);
+
static struct attribute *qat_attrs[] = {
&dev_attr_state.attr,
&dev_attr_cfg_services.attr,
&dev_attr_pm_idle_enabled.attr,
&dev_attr_rp2srv.attr,
+ &dev_attr_num_rps.attr,
NULL,
};

--
2.41.0

2023-10-11 12:36:53

by Damian Muszynski

[permalink] [raw]
Subject: [PATCH 09/11] crypto: qat - add rate limiting sysfs interface

From: Ciunas Bennett <[email protected]>

Add an interface for the rate limiting feature which allows to add,
remove and modify a QAT SLA (Service Level Agreement).

This adds a new sysfs attribute group, `qat_rl`, which can be accessed
from /sys/bus/pci/devices/<BUS:DEV:FUNCTION> with the following
hierarchy:
|-+ qat_rl
|---- id (RW) # SLA identifier
|---- cir (RW) # Committed Information Rate
|---- pir (RW) # Peak Information Rate
|---- srv (RW) # Service to be rate limited
|---- rp (RW) (HEX) # Ring pairs to be rate limited
|---- cap_rem (RW) # Remaining capability for a service
|---- sla_op (WO) # Allows to perform an operation on an SLA

The API works by setting the appropriate RW attributes and then
issuing a command through the `sla_op`. For example, to create an SLA, a
user needs to input the necessary data into the attributes cir, pir, srv
and rp and then write into `sla_op` the command `add` to execute the
operation.
The API also provides `cap_rem` attribute to get information about
the remaining device capability within a certain service which is
required when setting an SLA.

Signed-off-by: Ciunas Bennett <[email protected]>
Reviewed-by: Giovanni Cabiddu <[email protected]>
Reviewed-by: Damian Muszynski <[email protected]>
---
Documentation/ABI/testing/sysfs-driver-qat_rl | 227 +++++++++
drivers/crypto/intel/qat/qat_common/Makefile | 1 +
drivers/crypto/intel/qat/qat_common/adf_rl.c | 10 +
drivers/crypto/intel/qat/qat_common/adf_rl.h | 7 +
.../intel/qat/qat_common/adf_sysfs_rl.c | 451 ++++++++++++++++++
.../intel/qat/qat_common/adf_sysfs_rl.h | 11 +
6 files changed, 707 insertions(+)
create mode 100644 Documentation/ABI/testing/sysfs-driver-qat_rl
create mode 100644 drivers/crypto/intel/qat/qat_common/adf_sysfs_rl.c
create mode 100644 drivers/crypto/intel/qat/qat_common/adf_sysfs_rl.h

diff --git a/Documentation/ABI/testing/sysfs-driver-qat_rl b/Documentation/ABI/testing/sysfs-driver-qat_rl
new file mode 100644
index 000000000000..41e774cc8646
--- /dev/null
+++ b/Documentation/ABI/testing/sysfs-driver-qat_rl
@@ -0,0 +1,227 @@
+What: /sys/bus/pci/devices/<BDF>/qat_rl/sla_op
+Date: January 2024
+KernelVersion: 6.7
+Contact: [email protected]
+Description:
+ (WO) This attribute is used to perform an operation on an SLA.
+ The supported operations are: add, update, rm, rm_all, and get.
+
+ Input values must be filled through the associated attribute in
+ this group before a write to this file.
+ If the operation completes successfully, the associated
+ attributes will be updated.
+ The associated attributes are: cir, pir, srv, rp, and id.
+
+ Supported operations:
+
+ * add: Creates a new SLA with the provided inputs from user.
+ * Inputs: cir, pir, srv, and rp
+ * Output: id
+
+ * get: Returns the configuration of the specified SLA in id attribute
+ * Inputs: id
+ * Outputs: cir, pir, srv, and rp
+
+ * update: Updates the SLA with new values set in the following attributes
+ * Inputs: id, cir, and pir
+
+ * rm: Removes the specified SLA in the id attribute.
+ * Inputs: id
+
+ * rm_all: Removes all the configured SLAs.
+ * Inputs: None
+
+ This attribute is only available for qat_4xxx devices.
+
+What: /sys/bus/pci/devices/<BDF>/qat_rl/rp
+Date: January 2024
+KernelVersion: 6.7
+Contact: [email protected]
+Description:
+ (RW) When read, reports the current assigned ring pairs for the
+ queried SLA.
+ When wrote to, configures the ring pairs associated to a new SLA.
+
+ The value is a 64-bit bit mask and is written/displayed in hex.
+ Each bit of this mask represents a single ring pair i.e.,
+ bit 1 == ring pair id 0; bit 3 == ring pair id 2.
+
+ Selected ring pairs must to be assigned to a single service,
+ i.e. the one provided with the srv attribute. The service
+ assigned to a certain ring pair can be checked by querying
+ the attribute qat/rp2srv.
+
+ The maximum number of ring pairs is 4 per SLA.
+
+ Applicability in sla_op:
+
+ * WRITE: add operation
+ * READ: get operation
+
+ Example usage::
+
+ ## Read
+ # echo 4 > /sys/bus/pci/devices/<BDF>/qat_rl/id
+ # cat /sys/bus/pci/devices/<BDF>/qat_rl/rp
+ 0x5
+
+ ## Write
+ # echo 0x5 > /sys/bus/pci/devices/<BDF>/qat_rl/rp
+
+ This attribute is only available for qat_4xxx devices.
+
+What: /sys/bus/pci/devices/<BDF>/qat_rl/id
+Date: January 2024
+KernelVersion: 6.7
+Contact: [email protected]
+Description:
+ (RW) If written to, the value is used to retrieve a particular
+ SLA and operate on it.
+ This is valid only for the following operations: update, rm,
+ and get.
+ A read of this attribute is only guaranteed to have correct data
+ after creation of an SLA.
+
+ Applicability in sla_op:
+
+ * WRITE: rm and update operations
+ * READ: add and get operations
+
+ Example usage::
+
+ ## Read
+ ## Set attributes e.g. cir, pir, srv, etc
+ # echo "add" > /sys/bus/pci/devices/<BDF>/qat_rl/sla_op
+ # cat /sys/bus/pci/devices/<BDF>/qat_rl/id
+ 4
+
+ ## Write
+ # echo 7 > /sys/bus/pci/devices/<BDF>/qat_rl/id
+ # echo "get" > /sys/bus/pci/devices/<BDF>/qat_rl/sla_op
+ # cat /sys/bus/pci/devices/<BDF>/qat_rl/rp
+ 0x5 ## ring pair ID 0 and ring pair ID 2
+
+ This attribute is only available for qat_4xxx devices.
+
+What: /sys/bus/pci/devices/<BDF>/qat_rl/cir
+Date: January 2024
+KernelVersion: 6.7
+Contact: [email protected]
+Description:
+ (RW) Committed information rate (CIR). Rate guaranteed to be
+ achieved by a particular SLA. The value is expressed in
+ permille scale, i.e. 1000 refers to the maximum device
+ throughput for a selected service.
+
+ After sending a "get" to sla_op, this will be populated with the
+ CIR for that queried SLA.
+ Write to this file before sending an "add/update" sla_op, to set
+ the SLA to the specified value.
+
+ Applicability in sla_op:
+
+ * WRITE: add and update operations
+ * READ: get operation
+
+ Example usage::
+
+ ## Write
+ # echo 500 > /sys/bus/pci/devices/<BDF>/qat_rl/cir
+ # echo "add" /sys/bus/pci/devices/<BDF>/qat_rl/sla_op
+
+ ## Read
+ # echo 4 > /sys/bus/pci/devices/<BDF>/qat_rl/id
+ # echo "get" > /sys/bus/pci/devices/<BDF>/qat_rl/sla_op
+ # cat /sys/bus/pci/devices/<BDF>/qat_rl/cir
+ 500
+
+ This attribute is only available for qat_4xxx devices.
+
+What: /sys/bus/pci/devices/<BDF>/qat_rl/pir
+Date: January 2024
+KernelVersion: 6.7
+Contact: [email protected]
+Description:
+ (RW) Peak information rate (PIR). The maximum rate that can be
+ achieved by that particular SLA. An SLA can reach a value
+ between CIR and PIR when the device is not fully utilized by
+ requests from other users (assigned to different SLAs).
+
+ After sending a "get" to sla_op, this will be populated with the
+ PIR for that queried SLA.
+ Write to this file before sending an "add/update" sla_op, to set
+ the SLA to the specified value.
+
+ Applicability in sla_op:
+
+ * WRITE: add and update operations
+ * READ: get operation
+
+ Example usage::
+
+ ## Write
+ # echo 750 > /sys/bus/pci/devices/<BDF>/qat_rl/pir
+ # echo "add" > /sys/bus/pci/devices/<BDF>/qat_rl/sla_op
+
+ ## Read
+ # echo 4 > /sys/bus/pci/devices/<BDF>/qat_rl/id
+ # echo "get" > /sys/bus/pci/devices/<BDF>/qat_rl/sla_op
+ # cat /sys/bus/pci/devices/<BDF>/qat_rl/pir
+ 750
+
+ This attribute is only available for qat_4xxx devices.
+
+What: /sys/bus/pci/devices/<BDF>/qat_rl/srv
+Date: January 2024
+KernelVersion: 6.7
+Contact: [email protected]
+Description:
+ (RW) Service (SRV). Represents the service (sym, asym, dc)
+ associated to an SLA.
+ Can be written to or queried to set/show the SRV type for an SLA.
+ The SRV attribute is used to specify the SRV type before adding
+ an SLA. After an SLA is configured, reports the service
+ associated to that SLA.
+
+ Applicability in sla_op:
+
+ * WRITE: add and update operations
+ * READ: get operation
+
+ Example usage::
+
+ ## Write
+ # echo "dc" > /sys/bus/pci/devices/<BDF>/qat_rl/srv
+ # echo "add" > /sys/bus/pci/devices/<BDF>/qat_rl/sla_op
+ # cat /sys/bus/pci/devices/<BDF>/qat_rl/id
+ 4
+
+ ## Read
+ # echo 4 > /sys/bus/pci/devices/<BDF>/qat_rl/id
+ # echo "get" > /sys/bus/pci/devices/<BDF>/qat_rl/sla_op
+ # cat /sys/bus/pci/devices/<BDF>/qat_rl/srv
+ dc
+
+ This attribute is only available for qat_4xxx devices.
+
+What: /sys/bus/pci/devices/<BDF>/qat_rl/cap_rem
+Date: January 2024
+KernelVersion: 6.7
+Contact: [email protected]
+Description:
+ (RW) This file will return the remaining capability for a
+ particular service/sla. This is the remaining value that a new
+ SLA can be set to or a current SLA can be increased with.
+
+ Example usage::
+
+ # echo "asym" > /sys/bus/pci/devices/<BDF>/qat_rl/cap_rem
+ # cat /sys/bus/pci/devices/<BDF>/qat_rl/cap_rem
+ 250
+ # echo 250 > /sys/bus/pci/devices/<BDF>/qat_rl/cir
+ # echo "add" > /sys/bus/pci/devices/<BDF>/qat_rl/sla_op
+ # cat /sys/bus/pci/devices/<BDF>/qat_rl/cap_rem
+ 0
+
+ This attribute is only available for qat_4xxx devices.
+
diff --git a/drivers/crypto/intel/qat/qat_common/Makefile b/drivers/crypto/intel/qat/qat_common/Makefile
index 77f8c484d49c..832f0b48d99e 100644
--- a/drivers/crypto/intel/qat/qat_common/Makefile
+++ b/drivers/crypto/intel/qat/qat_common/Makefile
@@ -28,6 +28,7 @@ intel_qat-objs := adf_cfg.o \
qat_algs_send.o \
adf_rl.o \
adf_rl_admin.o \
+ adf_sysfs_rl.o \
qat_uclo.o \
qat_hal.o \
qat_bl.o
diff --git a/drivers/crypto/intel/qat/qat_common/adf_rl.c b/drivers/crypto/intel/qat/qat_common/adf_rl.c
index dec98a056dce..d8bb4154048a 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_rl.c
+++ b/drivers/crypto/intel/qat/qat_common/adf_rl.c
@@ -16,6 +16,7 @@
#include "adf_common_drv.h"
#include "adf_rl_admin.h"
#include "adf_rl.h"
+#include "adf_sysfs_rl.h"

#define RL_TOKEN_GRANULARITY_PCIEIN_BUCKET 0U
#define RL_TOKEN_GRANULARITY_PCIEOUT_BUCKET 0U
@@ -1147,8 +1148,16 @@ int adf_rl_start(struct adf_accel_dev *accel_dev)
goto ret_sla_rm;
}

+ ret = adf_sysfs_rl_add(accel_dev);
+ if (ret) {
+ dev_err(&GET_DEV(accel_dev), "failed to add sysfs interface\n");
+ goto ret_sysfs_rm;
+ }
+
return 0;

+ret_sysfs_rm:
+ adf_sysfs_rl_rm(accel_dev);
ret_sla_rm:
adf_rl_remove_sla_all(accel_dev, true);
ret_free:
@@ -1163,6 +1172,7 @@ void adf_rl_stop(struct adf_accel_dev *accel_dev)
if (!accel_dev->rate_limiting)
return;

+ adf_sysfs_rl_rm(accel_dev);
adf_rl_remove_sla_all(accel_dev, true);
}

diff --git a/drivers/crypto/intel/qat/qat_common/adf_rl.h b/drivers/crypto/intel/qat/qat_common/adf_rl.h
index 1ccb6613c92e..eb5a330f8543 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_rl.h
+++ b/drivers/crypto/intel/qat/qat_common/adf_rl.h
@@ -75,6 +75,12 @@ struct rl_slice_cnt {
u8 cph_cnt;
};

+struct adf_rl_interface_data {
+ struct adf_rl_sla_input_data input;
+ enum adf_base_services cap_rem_srv;
+ struct rw_semaphore lock;
+};
+
struct adf_rl_hw_data {
u32 scale_ref;
u32 scan_interval;
@@ -113,6 +119,7 @@ struct adf_rl {
bool rp_in_use[RL_RP_CNT_MAX];
/* Mutex protecting writing to SLAs lists */
struct mutex rl_lock;
+ struct adf_rl_interface_data user_input;
};

/**
diff --git a/drivers/crypto/intel/qat/qat_common/adf_sysfs_rl.c b/drivers/crypto/intel/qat/qat_common/adf_sysfs_rl.c
new file mode 100644
index 000000000000..abf9c52474ec
--- /dev/null
+++ b/drivers/crypto/intel/qat/qat_common/adf_sysfs_rl.c
@@ -0,0 +1,451 @@
+// SPDX-License-Identifier: GPL-2.0-only
+/* Copyright(c) 2023 Intel Corporation */
+
+#define dev_fmt(fmt) "RateLimiting: " fmt
+
+#include <linux/dev_printk.h>
+#include <linux/pci.h>
+#include <linux/sysfs.h>
+#include <linux/types.h>
+
+#include "adf_common_drv.h"
+#include "adf_rl.h"
+#include "adf_sysfs_rl.h"
+
+#define GET_RL_STRUCT(accel_dev) ((accel_dev)->rate_limiting->user_input)
+
+enum rl_ops {
+ ADD,
+ UPDATE,
+ RM,
+ RM_ALL,
+ GET,
+};
+
+enum rl_params {
+ RP_MASK,
+ ID,
+ CIR,
+ PIR,
+ SRV,
+ CAP_REM_SRV,
+};
+
+static const char *const rl_services[] = {
+ [ADF_SVC_ASYM] = "asym",
+ [ADF_SVC_SYM] = "sym",
+ [ADF_SVC_DC] = "dc",
+};
+
+static const char *const rl_operations[] = {
+ [ADD] = "add",
+ [UPDATE] = "update",
+ [RM] = "rm",
+ [RM_ALL] = "rm_all",
+ [GET] = "get",
+};
+
+static int set_param_u(struct device *dev, enum rl_params param, u64 set)
+{
+ struct adf_rl_interface_data *data;
+ struct adf_accel_dev *accel_dev;
+ int ret = 0;
+
+ accel_dev = adf_devmgr_pci_to_accel_dev(to_pci_dev(dev));
+ if (!accel_dev)
+ return -EINVAL;
+
+ data = &GET_RL_STRUCT(accel_dev);
+
+ down_write(&data->lock);
+ switch (param) {
+ case RP_MASK:
+ data->input.rp_mask = set;
+ break;
+ case CIR:
+ data->input.cir = set;
+ break;
+ case PIR:
+ data->input.pir = set;
+ break;
+ case SRV:
+ data->input.srv = set;
+ break;
+ case CAP_REM_SRV:
+ data->cap_rem_srv = set;
+ break;
+ default:
+ ret = -EINVAL;
+ break;
+ }
+ up_write(&data->lock);
+
+ return ret;
+}
+
+static int set_param_s(struct device *dev, enum rl_params param, int set)
+{
+ struct adf_rl_interface_data *data;
+ struct adf_accel_dev *accel_dev;
+
+ accel_dev = adf_devmgr_pci_to_accel_dev(to_pci_dev(dev));
+ if (!accel_dev || param != ID)
+ return -EINVAL;
+
+ data = &GET_RL_STRUCT(accel_dev);
+
+ down_write(&data->lock);
+ data->input.sla_id = set;
+ up_write(&data->lock);
+
+ return 0;
+}
+
+static int get_param_u(struct device *dev, enum rl_params param, u64 *get)
+{
+ struct adf_rl_interface_data *data;
+ struct adf_accel_dev *accel_dev;
+ int ret = 0;
+
+ accel_dev = adf_devmgr_pci_to_accel_dev(to_pci_dev(dev));
+ if (!accel_dev)
+ return -EINVAL;
+
+ data = &GET_RL_STRUCT(accel_dev);
+
+ down_read(&data->lock);
+ switch (param) {
+ case RP_MASK:
+ *get = data->input.rp_mask;
+ break;
+ case CIR:
+ *get = data->input.cir;
+ break;
+ case PIR:
+ *get = data->input.pir;
+ break;
+ case SRV:
+ *get = data->input.srv;
+ break;
+ default:
+ ret = -EINVAL;
+ }
+ up_read(&data->lock);
+
+ return ret;
+}
+
+static int get_param_s(struct device *dev, enum rl_params param)
+{
+ struct adf_rl_interface_data *data;
+ struct adf_accel_dev *accel_dev;
+ int ret = 0;
+
+ accel_dev = adf_devmgr_pci_to_accel_dev(to_pci_dev(dev));
+ if (!accel_dev)
+ return -EINVAL;
+
+ data = &GET_RL_STRUCT(accel_dev);
+
+ down_read(&data->lock);
+ if (param == ID)
+ ret = data->input.sla_id;
+ up_read(&data->lock);
+
+ return ret;
+}
+
+static ssize_t rp_show(struct device *dev, struct device_attribute *attr,
+ char *buf)
+{
+ int ret;
+ u64 get;
+
+ ret = get_param_u(dev, RP_MASK, &get);
+ if (ret)
+ return ret;
+
+ return sysfs_emit(buf, "%#llx\n", get);
+}
+
+static ssize_t rp_store(struct device *dev, struct device_attribute *attr,
+ const char *buf, size_t count)
+{
+ int err;
+ u64 val;
+
+ err = kstrtou64(buf, 16, &val);
+ if (err)
+ return err;
+
+ err = set_param_u(dev, RP_MASK, val);
+ if (err)
+ return err;
+
+ return count;
+}
+static DEVICE_ATTR_RW(rp);
+
+static ssize_t id_show(struct device *dev, struct device_attribute *attr,
+ char *buf)
+{
+ return sysfs_emit(buf, "%d\n", get_param_s(dev, ID));
+}
+
+static ssize_t id_store(struct device *dev, struct device_attribute *attr,
+ const char *buf, size_t count)
+{
+ int err;
+ int val;
+
+ err = kstrtoint(buf, 10, &val);
+ if (err)
+ return err;
+
+ err = set_param_s(dev, ID, val);
+ if (err)
+ return err;
+
+ return count;
+}
+static DEVICE_ATTR_RW(id);
+
+static ssize_t cir_show(struct device *dev, struct device_attribute *attr,
+ char *buf)
+{
+ int ret;
+ u64 get;
+
+ ret = get_param_u(dev, CIR, &get);
+ if (ret)
+ return ret;
+
+ return sysfs_emit(buf, "%llu\n", get);
+}
+
+static ssize_t cir_store(struct device *dev, struct device_attribute *attr,
+ const char *buf, size_t count)
+{
+ unsigned int val;
+ int err;
+
+ err = kstrtouint(buf, 10, &val);
+ if (err)
+ return err;
+
+ err = set_param_u(dev, CIR, val);
+ if (err)
+ return err;
+
+ return count;
+}
+static DEVICE_ATTR_RW(cir);
+
+static ssize_t pir_show(struct device *dev, struct device_attribute *attr,
+ char *buf)
+{
+ int ret;
+ u64 get;
+
+ ret = get_param_u(dev, PIR, &get);
+ if (ret)
+ return ret;
+
+ return sysfs_emit(buf, "%llu\n", get);
+}
+
+static ssize_t pir_store(struct device *dev, struct device_attribute *attr,
+ const char *buf, size_t count)
+{
+ unsigned int val;
+ int err;
+
+ err = kstrtouint(buf, 10, &val);
+ if (err)
+ return err;
+
+ err = set_param_u(dev, PIR, val);
+ if (err)
+ return err;
+
+ return count;
+}
+static DEVICE_ATTR_RW(pir);
+
+static ssize_t srv_show(struct device *dev, struct device_attribute *attr,
+ char *buf)
+{
+ int ret;
+ u64 get;
+
+ ret = get_param_u(dev, SRV, &get);
+ if (ret)
+ return ret;
+
+ if (get == ADF_SVC_NONE)
+ return -EINVAL;
+
+ return sysfs_emit(buf, "%s\n", rl_services[get]);
+}
+
+static ssize_t srv_store(struct device *dev, struct device_attribute *attr,
+ const char *buf, size_t count)
+{
+ unsigned int val;
+ int ret;
+
+ ret = sysfs_match_string(rl_services, buf);
+ if (ret < 0)
+ return ret;
+
+ val = ret;
+ ret = set_param_u(dev, SRV, val);
+ if (ret)
+ return ret;
+
+ return count;
+}
+static DEVICE_ATTR_RW(srv);
+
+static ssize_t cap_rem_show(struct device *dev, struct device_attribute *attr,
+ char *buf)
+{
+ struct adf_rl_interface_data *data;
+ struct adf_accel_dev *accel_dev;
+ int ret, rem_cap;
+
+ accel_dev = adf_devmgr_pci_to_accel_dev(to_pci_dev(dev));
+ if (!accel_dev)
+ return -EINVAL;
+
+ data = &GET_RL_STRUCT(accel_dev);
+
+ down_read(&data->lock);
+ rem_cap = adf_rl_get_capability_remaining(accel_dev, data->cap_rem_srv,
+ RL_SLA_EMPTY_ID);
+ up_read(&data->lock);
+ if (rem_cap < 0)
+ return rem_cap;
+
+ ret = sysfs_emit(buf, "%u\n", rem_cap);
+
+ return ret;
+}
+
+static ssize_t cap_rem_store(struct device *dev, struct device_attribute *attr,
+ const char *buf, size_t count)
+{
+ unsigned int val;
+ int ret;
+
+ ret = sysfs_match_string(rl_services, buf);
+ if (ret < 0)
+ return ret;
+
+ val = ret;
+ ret = set_param_u(dev, CAP_REM_SRV, val);
+ if (ret)
+ return ret;
+
+ return count;
+}
+static DEVICE_ATTR_RW(cap_rem);
+
+static ssize_t sla_op_store(struct device *dev, struct device_attribute *attr,
+ const char *buf, size_t count)
+{
+ struct adf_rl_interface_data *data;
+ struct adf_accel_dev *accel_dev;
+ int ret;
+
+ accel_dev = adf_devmgr_pci_to_accel_dev(to_pci_dev(dev));
+ if (!accel_dev)
+ return -EINVAL;
+
+ data = &GET_RL_STRUCT(accel_dev);
+
+ ret = sysfs_match_string(rl_operations, buf);
+ if (ret < 0)
+ return ret;
+
+ down_write(&data->lock);
+ switch (ret) {
+ case ADD:
+ data->input.parent_id = RL_PARENT_DEFAULT_ID;
+ data->input.type = RL_LEAF;
+ data->input.sla_id = 0;
+ ret = adf_rl_add_sla(accel_dev, &data->input);
+ if (ret)
+ goto err_free_lock;
+ break;
+ case UPDATE:
+ ret = adf_rl_update_sla(accel_dev, &data->input);
+ if (ret)
+ goto err_free_lock;
+ break;
+ case RM:
+ ret = adf_rl_remove_sla(accel_dev, data->input.sla_id);
+ if (ret)
+ goto err_free_lock;
+ break;
+ case RM_ALL:
+ adf_rl_remove_sla_all(accel_dev, false);
+ break;
+ case GET:
+ ret = adf_rl_get_sla(accel_dev, &data->input);
+ if (ret)
+ goto err_free_lock;
+ break;
+ default:
+ ret = -EINVAL;
+ goto err_free_lock;
+ }
+ up_write(&data->lock);
+
+ return count;
+
+err_free_lock:
+ up_write(&data->lock);
+
+ return ret;
+}
+static DEVICE_ATTR_WO(sla_op);
+
+static struct attribute *qat_rl_attrs[] = {
+ &dev_attr_rp.attr,
+ &dev_attr_id.attr,
+ &dev_attr_cir.attr,
+ &dev_attr_pir.attr,
+ &dev_attr_srv.attr,
+ &dev_attr_cap_rem.attr,
+ &dev_attr_sla_op.attr,
+ NULL,
+};
+
+static struct attribute_group qat_rl_group = {
+ .attrs = qat_rl_attrs,
+ .name = "qat_rl",
+};
+
+int adf_sysfs_rl_add(struct adf_accel_dev *accel_dev)
+{
+ struct adf_rl_interface_data *data;
+ int ret;
+
+ data = &GET_RL_STRUCT(accel_dev);
+
+ ret = device_add_group(&GET_DEV(accel_dev), &qat_rl_group);
+ if (ret)
+ dev_err(&GET_DEV(accel_dev),
+ "Failed to create qat_rl attribute group\n");
+
+ data->cap_rem_srv = ADF_SVC_NONE;
+ data->input.srv = ADF_SVC_NONE;
+
+ return ret;
+}
+
+void adf_sysfs_rl_rm(struct adf_accel_dev *accel_dev)
+{
+ device_remove_group(&GET_DEV(accel_dev), &qat_rl_group);
+}
diff --git a/drivers/crypto/intel/qat/qat_common/adf_sysfs_rl.h b/drivers/crypto/intel/qat/qat_common/adf_sysfs_rl.h
new file mode 100644
index 000000000000..22d36aa8a757
--- /dev/null
+++ b/drivers/crypto/intel/qat/qat_common/adf_sysfs_rl.h
@@ -0,0 +1,11 @@
+/* SPDX-License-Identifier: GPL-2.0-only */
+/* Copyright(c) 2023 Intel Corporation */
+#ifndef ADF_SYSFS_RL_H_
+#define ADF_SYSFS_RL_H_
+
+struct adf_accel_dev;
+
+int adf_sysfs_rl_add(struct adf_accel_dev *accel_dev);
+void adf_sysfs_rl_rm(struct adf_accel_dev *accel_dev);
+
+#endif /* ADF_SYSFS_RL_H_ */
--
2.41.0

2023-10-11 12:36:56

by Damian Muszynski

[permalink] [raw]
Subject: [PATCH 10/11] crypto: qat - add rp2svc sysfs attribute

From: Ciunas Bennett <[email protected]>

Add the attribute `rp2svc` to the `qat` attribute group. This provides a
way for a user to query a specific ring pair for the type of service
that is currently configured for.

When read, the service will be returned for the defined ring pair.
When written to this value will be stored as the ring pair to return
the service of.

Signed-off-by: Ciunas Bennett <[email protected]>
Reviewed-by: Giovanni Cabiddu <[email protected]>
Reviewed-by: Damian Muszynski <[email protected]>
---
Documentation/ABI/testing/sysfs-driver-qat | 32 +++++++++
.../intel/qat/qat_common/adf_accel_devices.h | 6 ++
.../crypto/intel/qat/qat_common/adf_sysfs.c | 66 +++++++++++++++++++
3 files changed, 104 insertions(+)

diff --git a/Documentation/ABI/testing/sysfs-driver-qat b/Documentation/ABI/testing/sysfs-driver-qat
index 96834d103a09..f24a5ddca94b 100644
--- a/Documentation/ABI/testing/sysfs-driver-qat
+++ b/Documentation/ABI/testing/sysfs-driver-qat
@@ -95,3 +95,35 @@ Description: (RW) This configuration option provides a way to force the device i
0

This attribute is only available for qat_4xxx devices.
+
+What: /sys/bus/pci/devices/<BDF>/qat/rp2srv
+Date: January 2024
+KernelVersion: 6.7
+Contact: [email protected]
+Description:
+ (RW) This attribute provides a way for a user to query a
+ specific ring pair for the type of service that it is currently
+ configured for.
+
+ When written to, the value is cached and used to perform the
+ read operation. Allowed values are in the range 0 to N-1, where
+ N is the max number of ring pairs supported by a device. This
+ can be queried using the attribute qat/num_rps.
+
+ A read returns the service associated to the ring pair queried.
+
+ The values are:
+
+ * dc: the ring pair is configured for running compression services
+ * sym: the ring pair is configured for running symmetric crypto
+ services
+ * asym: the ring pair is configured for running asymmetric crypto
+ services
+
+ Example usage::
+
+ # echo 1 > /sys/bus/pci/devices/<BDF>/qat/rp2srv
+ # cat /sys/bus/pci/devices/<BDF>/qat/rp2srv
+ sym
+
+ This attribute is only available for qat_4xxx devices.
diff --git a/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h b/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
index 35b805b9a136..9bb9e30eb456 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
+++ b/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
@@ -307,6 +307,11 @@ struct adf_pm {
char __user *buf, size_t count, loff_t *pos);
};

+struct adf_sysfs {
+ int ring_num;
+ struct rw_semaphore lock; /* protects access to the fields in this struct */
+};
+
struct adf_accel_dev {
struct adf_etr_data *transport;
struct adf_hw_device_data *hw_device;
@@ -328,6 +333,7 @@ struct adf_accel_dev {
struct adf_timer *timer;
struct adf_heartbeat *heartbeat;
struct adf_rl *rate_limiting;
+ struct adf_sysfs sysfs;
union {
struct {
/* protects VF2PF interrupts access */
diff --git a/drivers/crypto/intel/qat/qat_common/adf_sysfs.c b/drivers/crypto/intel/qat/qat_common/adf_sysfs.c
index f4a89f7ed4e9..9317127128a9 100644
--- a/drivers/crypto/intel/qat/qat_common/adf_sysfs.c
+++ b/drivers/crypto/intel/qat/qat_common/adf_sysfs.c
@@ -8,6 +8,8 @@
#include "adf_cfg_services.h"
#include "adf_common_drv.h"

+#define UNSET_RING_NUM -1
+
static const char * const state_operations[] = {
[DEV_DOWN] = "down",
[DEV_UP] = "up",
@@ -205,10 +207,72 @@ static DEVICE_ATTR_RW(pm_idle_enabled);
static DEVICE_ATTR_RW(state);
static DEVICE_ATTR_RW(cfg_services);

+static ssize_t rp2srv_show(struct device *dev, struct device_attribute *attr,
+ char *buf)
+{
+ struct adf_hw_device_data *hw_data;
+ struct adf_accel_dev *accel_dev;
+ enum adf_cfg_service_type svc;
+
+ accel_dev = adf_devmgr_pci_to_accel_dev(to_pci_dev(dev));
+ hw_data = GET_HW_DATA(accel_dev);
+
+ if (accel_dev->sysfs.ring_num == UNSET_RING_NUM)
+ return -EINVAL;
+
+ down_read(&accel_dev->sysfs.lock);
+ svc = GET_SRV_TYPE(accel_dev, accel_dev->sysfs.ring_num %
+ hw_data->num_banks_per_vf);
+ up_read(&accel_dev->sysfs.lock);
+
+ switch (svc) {
+ case COMP:
+ return sysfs_emit(buf, "%s\n", ADF_CFG_DC);
+ case SYM:
+ return sysfs_emit(buf, "%s\n", ADF_CFG_SYM);
+ case ASYM:
+ return sysfs_emit(buf, "%s\n", ADF_CFG_ASYM);
+ default:
+ break;
+ }
+ return -EINVAL;
+}
+
+static ssize_t rp2srv_store(struct device *dev, struct device_attribute *attr,
+ const char *buf, size_t count)
+{
+ struct adf_accel_dev *accel_dev;
+ int ring, num_rings, ret;
+
+ accel_dev = adf_devmgr_pci_to_accel_dev(to_pci_dev(dev));
+ if (!accel_dev)
+ return -EINVAL;
+
+ ret = kstrtouint(buf, 10, &ring);
+ if (ret)
+ return ret;
+
+ num_rings = GET_MAX_BANKS(accel_dev);
+ if (ring >= num_rings) {
+ dev_err(&GET_DEV(accel_dev),
+ "Device does not support more than %u ring pairs\n",
+ num_rings);
+ return -EINVAL;
+ }
+
+ down_write(&accel_dev->sysfs.lock);
+ accel_dev->sysfs.ring_num = ring;
+ up_write(&accel_dev->sysfs.lock);
+
+ return count;
+}
+static DEVICE_ATTR_RW(rp2srv);
+
static struct attribute *qat_attrs[] = {
&dev_attr_state.attr,
&dev_attr_cfg_services.attr,
&dev_attr_pm_idle_enabled.attr,
+ &dev_attr_rp2srv.attr,
NULL,
};

@@ -227,6 +291,8 @@ int adf_sysfs_init(struct adf_accel_dev *accel_dev)
"Failed to create qat attribute group: %d\n", ret);
}

+ accel_dev->sysfs.ring_num = UNSET_RING_NUM;
+
return ret;
}
EXPORT_SYMBOL_GPL(adf_sysfs_init);
--
2.41.0

2023-10-16 07:23:23

by Tero Kristo

[permalink] [raw]
Subject: Re: [PATCH 07/11] crypto: qat - add retrieval of fw capabilities

Hello,

On 11/10/2023 15:15, Muszynski, Damian wrote:
> The QAT firmware provides a mechanism to retrieve its capabilities
> through the init admin interface.
>
> Add logic to retrieve the firmware capability mask from the firmware
> through the init/admin channel. This mask reports if the
> power management, telemetry and rate limiting features are supported.
>
> The fw capabilities are stored in the accel_dev structure and are used
> to detect if a certain feature is supported by the firmware loaded
> in the device.
>
> This is supported only by devices which have an admin AE.
>
> Signed-off-by: Damian Muszynski <[email protected]>
> Reviewed-by: Giovanni Cabiddu <[email protected]>
> ---
> .../intel/qat/qat_common/adf_accel_devices.h | 1 +
> .../crypto/intel/qat/qat_common/adf_admin.c | 25 +++++++++++++++++++
> .../qat/qat_common/icp_qat_fw_init_admin.h | 3 +++
> 3 files changed, 29 insertions(+)
>
> diff --git a/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h b/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
> index 3674904d0527..45742226a96f 100644
> --- a/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
> +++ b/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
> @@ -221,6 +221,7 @@ struct adf_hw_device_data {
> u32 straps;
> u32 accel_capabilities_mask;
> u32 extended_dc_capabilities;
> + u16 fw_capabilities;
> u32 clock_frequency;
> u32 instance_id;
> u16 accel_mask;
> diff --git a/drivers/crypto/intel/qat/qat_common/adf_admin.c b/drivers/crypto/intel/qat/qat_common/adf_admin.c
> index 15ffda582334..9ff00eb4cc67 100644
> --- a/drivers/crypto/intel/qat/qat_common/adf_admin.c
> +++ b/drivers/crypto/intel/qat/qat_common/adf_admin.c
> @@ -310,6 +310,26 @@ static bool is_dcc_enabled(struct adf_accel_dev *accel_dev)
> return !strcmp(services, "dcc");
> }
>
> +static int adf_get_fw_capabilities(struct adf_accel_dev *accel_dev, u16 *caps)
> +{
> + u32 ae_mask = accel_dev->hw_device->admin_ae_mask;
> + struct icp_qat_fw_init_admin_resp resp = { };
> + struct icp_qat_fw_init_admin_req req = { };
> + int ret;
> +
> + if (!ae_mask)
> + return 0;
> +
> + req.cmd_id = ICP_QAT_FW_CAPABILITIES_GET;
> + ret = adf_send_admin(accel_dev, &req, &resp, ae_mask);
> + if (ret)
> + return ret;
> +
> + *caps = resp.fw_capabilities;
> +
> + return 0;
> +}
> +
> /**
> * adf_send_admin_init() - Function sends init message to FW
> * @accel_dev: Pointer to acceleration device.
> @@ -321,6 +341,7 @@ static bool is_dcc_enabled(struct adf_accel_dev *accel_dev)
> int adf_send_admin_init(struct adf_accel_dev *accel_dev)
> {
> u32 dc_capabilities = 0;
> + u16 fw_capabilities = 0;
> int ret;
>
> ret = adf_set_fw_constants(accel_dev);
> @@ -340,6 +361,10 @@ int adf_send_admin_init(struct adf_accel_dev *accel_dev)
> }
> accel_dev->hw_device->extended_dc_capabilities = dc_capabilities;
>
> + ret = adf_get_fw_capabilities(accel_dev, &fw_capabilities);

You could just pass the accel_dev->hw_device->fw_capabilities instead of
fw_capabilities here, ignore the return value, and get rid of the local
variable.

-Tero

> + if (!ret)
> + accel_dev->hw_device->fw_capabilities = fw_capabilities;
> +
> return adf_init_ae(accel_dev);
> }
> EXPORT_SYMBOL_GPL(adf_send_admin_init);
> diff --git a/drivers/crypto/intel/qat/qat_common/icp_qat_fw_init_admin.h b/drivers/crypto/intel/qat/qat_common/icp_qat_fw_init_admin.h
> index 9e5ce419d875..e4de9a30e0bd 100644
> --- a/drivers/crypto/intel/qat/qat_common/icp_qat_fw_init_admin.h
> +++ b/drivers/crypto/intel/qat/qat_common/icp_qat_fw_init_admin.h
> @@ -16,6 +16,7 @@ enum icp_qat_fw_init_admin_cmd_id {
> ICP_QAT_FW_HEARTBEAT_SYNC = 7,
> ICP_QAT_FW_HEARTBEAT_GET = 8,
> ICP_QAT_FW_COMP_CAPABILITY_GET = 9,
> + ICP_QAT_FW_CRYPTO_CAPABILITY_GET = 10,
> ICP_QAT_FW_DC_CHAIN_INIT = 11,
> ICP_QAT_FW_HEARTBEAT_TIMER_SET = 13,
> ICP_QAT_FW_TIMER_GET = 19,
> @@ -109,10 +110,12 @@ struct icp_qat_fw_init_admin_resp {
> __u32 unsuccessful_count;
> __u64 resrvd8;
> };
> + __u16 fw_capabilities;
> };
> } __packed;
>
> #define ICP_QAT_FW_SYNC ICP_QAT_FW_HEARTBEAT_SYNC
> +#define ICP_QAT_FW_CAPABILITIES_GET ICP_QAT_FW_CRYPTO_CAPABILITY_GET
>
> #define ICP_QAT_NUMBER_OF_PM_EVENTS 8
>
---------------------------------------------------------------------
Intel Finland Oy
Registered Address: PL 281, 00181 Helsinki
Business Identity Code: 0357606 - 4
Domiciled in Helsinki

This e-mail and any attachments may contain confidential material for
the sole use of the intended recipient(s). Any review or distribution
by others is strictly prohibited. If you are not the intended
recipient, please contact the sender and delete all copies.

2023-10-16 10:54:04

by Tero Kristo

[permalink] [raw]
Subject: Re: [PATCH 08/11] crypto: qat - add rate limiting feature to qat_4xxx

Hello,

On 11/10/2023 15:15, Muszynski, Damian wrote:
> The Rate Limiting (RL) feature allows to control the rate of requests
> that can be submitted on a ring pair (RP). This allows sharing a QAT
> device among multiple users while ensuring a guaranteed throughput.
>
> Configuration of RL is accomplished through entities called SLAs
> (Service Level Agreement). Each SLA object gets a unique identifier
> and defines the limitations for a single service across up to four
> ring pairs (RPs count allocated to a single VF).
>
> The rate is determined using two fields:
> * CIR (Committed Information Rate), i.e., the guaranteed rate.
> * PIR (Peak Information Rate), i.e., the maximum rate achievable
> when the device has available resources.
> The rate values are expressed in permille scale i.e. 0-1000.
> Ring pair selection is achieved by providing a 64-bit mask, where
> each bit corresponds to one of the ring pairs.
>
> This adds an interface and logic that allow to add, update, retrieve
> and remove an SLA.
>
> Signed-off-by: Damian Muszynski <[email protected]>
> Reviewed-by: Giovanni Cabiddu <[email protected]>
> ---
> .../intel/qat/qat_4xxx/adf_4xxx_hw_data.c | 20 +
> .../intel/qat/qat_4xxx/adf_4xxx_hw_data.h | 13 +-
> drivers/crypto/intel/qat/qat_common/Makefile | 2 +
> .../intel/qat/qat_common/adf_accel_devices.h | 3 +
> .../crypto/intel/qat/qat_common/adf_admin.c | 47 +
> .../crypto/intel/qat/qat_common/adf_admin.h | 8 +
> .../intel/qat/qat_common/adf_gen4_hw_data.h | 7 +
> .../crypto/intel/qat/qat_common/adf_init.c | 9 +
> drivers/crypto/intel/qat/qat_common/adf_rl.c | 1176 +++++++++++++++++
> drivers/crypto/intel/qat/qat_common/adf_rl.h | 169 +++
> .../intel/qat/qat_common/adf_rl_admin.c | 98 ++
> .../intel/qat/qat_common/adf_rl_admin.h | 18 +
> .../qat/qat_common/icp_qat_fw_init_admin.h | 38 +
> 13 files changed, 1607 insertions(+), 1 deletion(-)
> create mode 100644 drivers/crypto/intel/qat/qat_common/adf_rl.c
> create mode 100644 drivers/crypto/intel/qat/qat_common/adf_rl.h
> create mode 100644 drivers/crypto/intel/qat/qat_common/adf_rl_admin.c
> create mode 100644 drivers/crypto/intel/qat/qat_common/adf_rl_admin.h
>
> diff --git a/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c b/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
> index 8a80701b7791..bc3745f0018f 100644
> --- a/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
> +++ b/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.c
> @@ -342,6 +342,24 @@ static u32 get_heartbeat_clock(struct adf_hw_device_data *self)
> return ADF_4XXX_KPT_COUNTER_FREQ;
> }
>
> +static void adf_init_rl_data(struct adf_rl_hw_data *rl_data)
> +{
> + rl_data->pciout_tb_offset = ADF_GEN4_RL_TOKEN_PCIEOUT_BUCKET_OFFSET;
> + rl_data->pciin_tb_offset = ADF_GEN4_RL_TOKEN_PCIEIN_BUCKET_OFFSET;
> + rl_data->r2l_offset = ADF_GEN4_RL_R2L_OFFSET;
> + rl_data->l2c_offset = ADF_GEN4_RL_L2C_OFFSET;
> + rl_data->c2s_offset = ADF_GEN4_RL_C2S_OFFSET;
> +
> + rl_data->pcie_scale_div = ADF_4XXX_RL_PCIE_SCALE_FACTOR_DIV;
> + rl_data->pcie_scale_mul = ADF_4XXX_RL_PCIE_SCALE_FACTOR_MUL;
> + rl_data->dcpr_correction = ADF_4XXX_RL_DCPR_CORRECTION;
> + rl_data->max_tp[ADF_SVC_ASYM] = ADF_4XXX_RL_MAX_TP_ASYM;
> + rl_data->max_tp[ADF_SVC_SYM] = ADF_4XXX_RL_MAX_TP_SYM;
> + rl_data->max_tp[ADF_SVC_DC] = ADF_4XXX_RL_MAX_TP_DC;
> + rl_data->scan_interval = ADF_4XXX_RL_SCANS_PER_SEC;
> + rl_data->scale_ref = ADF_4XXX_RL_SLICE_REF;
> +}
> +
> static void adf_enable_error_correction(struct adf_accel_dev *accel_dev)
> {
> struct adf_bar *misc_bar = &GET_BARS(accel_dev)[ADF_4XXX_PMISC_BAR];
> @@ -583,10 +601,12 @@ void adf_init_hw_data_4xxx(struct adf_hw_device_data *hw_data, u32 dev_id)
> hw_data->stop_timer = adf_gen4_timer_stop;
> hw_data->get_hb_clock = get_heartbeat_clock;
> hw_data->num_hb_ctrs = ADF_NUM_HB_CNT_PER_AE;
> + hw_data->clock_frequency = ADF_4XXX_AE_FREQ;
>
> adf_gen4_init_hw_csr_ops(&hw_data->csr_ops);
> adf_gen4_init_pf_pfvf_ops(&hw_data->pfvf_ops);
> adf_gen4_init_dc_ops(&hw_data->dc_ops);
> + adf_init_rl_data(&hw_data->rl_data);
> }
>
> void adf_clean_hw_data_4xxx(struct adf_hw_device_data *hw_data)
> diff --git a/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.h b/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.h
> index bb3d95a8fb21..590cbe816b7b 100644
> --- a/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.h
> +++ b/drivers/crypto/intel/qat/qat_4xxx/adf_4xxx_hw_data.h
> @@ -65,8 +65,19 @@
> #define ADF_402XX_ASYM_OBJ "qat_402xx_asym.bin"
> #define ADF_402XX_ADMIN_OBJ "qat_402xx_admin.bin"
>
> +/* RL constants */
> +#define ADF_4XXX_RL_PCIE_SCALE_FACTOR_DIV 100
> +#define ADF_4XXX_RL_PCIE_SCALE_FACTOR_MUL 102
> +#define ADF_4XXX_RL_DCPR_CORRECTION 1
> +#define ADF_4XXX_RL_SCANS_PER_SEC 954
> +#define ADF_4XXX_RL_MAX_TP_ASYM 173750UL
> +#define ADF_4XXX_RL_MAX_TP_SYM 95000UL
> +#define ADF_4XXX_RL_MAX_TP_DC 45000UL
> +#define ADF_4XXX_RL_SLICE_REF 1000UL
> +
> /* Clocks frequency */
> -#define ADF_4XXX_KPT_COUNTER_FREQ (100 * HZ_PER_MHZ)
> +#define ADF_4XXX_KPT_COUNTER_FREQ (100 * HZ_PER_MHZ)
> +#define ADF_4XXX_AE_FREQ (1000 * HZ_PER_MHZ)
>
> /* qat_4xxx fuse bits are different from old GENs, redefine them */
> enum icp_qat_4xxx_slice_mask {
> diff --git a/drivers/crypto/intel/qat/qat_common/Makefile b/drivers/crypto/intel/qat/qat_common/Makefile
> index 204c7d0aa31e..77f8c484d49c 100644
> --- a/drivers/crypto/intel/qat/qat_common/Makefile
> +++ b/drivers/crypto/intel/qat/qat_common/Makefile
> @@ -26,6 +26,8 @@ intel_qat-objs := adf_cfg.o \
> qat_algs.o \
> qat_asym_algs.o \
> qat_algs_send.o \
> + adf_rl.o \
> + adf_rl_admin.o \
> qat_uclo.o \
> qat_hal.o \
> qat_bl.o
> diff --git a/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h b/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
> index 45742226a96f..35b805b9a136 100644
> --- a/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
> +++ b/drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
> @@ -8,6 +8,7 @@
> #include <linux/io.h>
> #include <linux/ratelimit.h>
> #include "adf_cfg_common.h"
> +#include "adf_rl.h"
> #include "adf_pfvf_msg.h"
>
> #define ADF_DH895XCC_DEVICE_NAME "dh895xcc"
> @@ -215,6 +216,7 @@ struct adf_hw_device_data {
> struct adf_pfvf_ops pfvf_ops;
> struct adf_hw_csr_ops csr_ops;
> struct adf_dc_ops dc_ops;
> + struct adf_rl_hw_data rl_data;
> const char *fw_name;
> const char *fw_mmp_name;
> u32 fuses;
> @@ -325,6 +327,7 @@ struct adf_accel_dev {
> struct adf_accel_pci accel_pci_dev;
> struct adf_timer *timer;
> struct adf_heartbeat *heartbeat;
> + struct adf_rl *rate_limiting;
> union {
> struct {
> /* protects VF2PF interrupts access */
> diff --git a/drivers/crypto/intel/qat/qat_common/adf_admin.c b/drivers/crypto/intel/qat/qat_common/adf_admin.c
> index 9ff00eb4cc67..b9cd21f772d3 100644
> --- a/drivers/crypto/intel/qat/qat_common/adf_admin.c
> +++ b/drivers/crypto/intel/qat/qat_common/adf_admin.c
> @@ -330,6 +330,53 @@ static int adf_get_fw_capabilities(struct adf_accel_dev *accel_dev, u16 *caps)
> return 0;
> }
>
> +int adf_send_admin_rl_init(struct adf_accel_dev *accel_dev,
> + struct icp_qat_fw_init_admin_slice_cnt *slices)
> +{
> + u32 ae_mask = accel_dev->hw_device->admin_ae_mask;
> + struct icp_qat_fw_init_admin_resp resp = { };
> + struct icp_qat_fw_init_admin_req req = { };
> + int ret;
> +
> + req.cmd_id = ICP_QAT_FW_RL_INIT;
> +
> + ret = adf_send_admin(accel_dev, &req, &resp, ae_mask);
> + if (ret)
> + return ret;
> +
> + memcpy(slices, &resp.slices, sizeof(*slices));
> +
> + return 0;
> +}
> +
> +int adf_send_admin_rl_add_update(struct adf_accel_dev *accel_dev,
> + struct icp_qat_fw_init_admin_req *req)
> +{
> + u32 ae_mask = accel_dev->hw_device->admin_ae_mask;
> + struct icp_qat_fw_init_admin_resp resp = { };
> +
> + /*
> + * req struct filled in rl implementation. Used commands
> + * ICP_QAT_FW_RL_ADD for a new SLA
> + * ICP_QAT_FW_RL_UPDATE for update SLA
> + */
> + return adf_send_admin(accel_dev, req, &resp, ae_mask);
> +}
> +
> +int adf_send_admin_rl_delete(struct adf_accel_dev *accel_dev, u16 node_id,
> + u8 node_type)
> +{
> + u32 ae_mask = accel_dev->hw_device->admin_ae_mask;
> + struct icp_qat_fw_init_admin_resp resp = { };
> + struct icp_qat_fw_init_admin_req req = { };
> +
> + req.cmd_id = ICP_QAT_FW_RL_REMOVE;
> + req.node_id = node_id;
> + req.node_type = node_type;
> +
> + return adf_send_admin(accel_dev, &req, &resp, ae_mask);
> +}
> +
> /**
> * adf_send_admin_init() - Function sends init message to FW
> * @accel_dev: Pointer to acceleration device.
> diff --git a/drivers/crypto/intel/qat/qat_common/adf_admin.h b/drivers/crypto/intel/qat/qat_common/adf_admin.h
> index 03507ec3a51d..55cbcbc66c9f 100644
> --- a/drivers/crypto/intel/qat/qat_common/adf_admin.h
> +++ b/drivers/crypto/intel/qat/qat_common/adf_admin.h
> @@ -3,6 +3,8 @@
> #ifndef ADF_ADMIN
> #define ADF_ADMIN
>
> +#include "icp_qat_fw_init_admin.h"
> +
> struct adf_accel_dev;
>
> int adf_init_admin_comms(struct adf_accel_dev *accel_dev);
> @@ -12,6 +14,12 @@ int adf_get_ae_fw_counters(struct adf_accel_dev *accel_dev, u16 ae, u64 *reqs, u
> int adf_init_admin_pm(struct adf_accel_dev *accel_dev, u32 idle_delay);
> int adf_send_admin_tim_sync(struct adf_accel_dev *accel_dev, u32 cnt);
> int adf_send_admin_hb_timer(struct adf_accel_dev *accel_dev, uint32_t ticks);
> +int adf_send_admin_rl_init(struct adf_accel_dev *accel_dev,
> + struct icp_qat_fw_init_admin_slice_cnt *slices);
> +int adf_send_admin_rl_add_update(struct adf_accel_dev *accel_dev,
> + struct icp_qat_fw_init_admin_req *req);
> +int adf_send_admin_rl_delete(struct adf_accel_dev *accel_dev, u16 node_id,
> + u8 node_type);
> int adf_get_fw_timestamp(struct adf_accel_dev *accel_dev, u64 *timestamp);
> int adf_get_pm_info(struct adf_accel_dev *accel_dev, dma_addr_t p_state_addr, size_t buff_size);
> int adf_get_cnv_stats(struct adf_accel_dev *accel_dev, u16 ae, u16 *err_cnt, u16 *latest_err);
> diff --git a/drivers/crypto/intel/qat/qat_common/adf_gen4_hw_data.h b/drivers/crypto/intel/qat/qat_common/adf_gen4_hw_data.h
> index 02d7a019ebf8..1813fe1d5a06 100644
> --- a/drivers/crypto/intel/qat/qat_common/adf_gen4_hw_data.h
> +++ b/drivers/crypto/intel/qat/qat_common/adf_gen4_hw_data.h
> @@ -139,6 +139,13 @@ do { \
> /* Number of heartbeat counter pairs */
> #define ADF_NUM_HB_CNT_PER_AE ADF_NUM_THREADS_PER_AE
>
> +/* Rate Limiting */
> +#define ADF_GEN4_RL_R2L_OFFSET 0x508000
> +#define ADF_GEN4_RL_L2C_OFFSET 0x509000
> +#define ADF_GEN4_RL_C2S_OFFSET 0x508818
> +#define ADF_GEN4_RL_TOKEN_PCIEIN_BUCKET_OFFSET 0x508800
> +#define ADF_GEN4_RL_TOKEN_PCIEOUT_BUCKET_OFFSET 0x508804
> +
> void adf_gen4_set_ssm_wdtimer(struct adf_accel_dev *accel_dev);
> void adf_gen4_init_hw_csr_ops(struct adf_hw_csr_ops *csr_ops);
> int adf_gen4_ring_pair_reset(struct adf_accel_dev *accel_dev, u32 bank_number);
> diff --git a/drivers/crypto/intel/qat/qat_common/adf_init.c b/drivers/crypto/intel/qat/qat_common/adf_init.c
> index b4cf605ccf3e..827c70c4221f 100644
> --- a/drivers/crypto/intel/qat/qat_common/adf_init.c
> +++ b/drivers/crypto/intel/qat/qat_common/adf_init.c
> @@ -9,6 +9,7 @@
> #include "adf_common_drv.h"
> #include "adf_dbgfs.h"
> #include "adf_heartbeat.h"
> +#include "adf_rl.h"
>
> static LIST_HEAD(service_table);
> static DEFINE_MUTEX(service_lock);
> @@ -133,6 +134,9 @@ static int adf_dev_init(struct adf_accel_dev *accel_dev)
> }
>
> adf_heartbeat_init(accel_dev);
> + ret = adf_rl_init(accel_dev);
> + if (ret && ret != -EOPNOTSUPP)
> + return ret;
>
> /*
> * Subservice initialisation is divided into two stages: init and start.
> @@ -208,6 +212,9 @@ static int adf_dev_start(struct adf_accel_dev *accel_dev)
> }
>
> adf_heartbeat_start(accel_dev);
> + ret = adf_rl_start(accel_dev);
> + if (ret && ret != -EOPNOTSUPP)
> + return ret;
>
> list_for_each_entry(service, &service_table, list) {
> if (service->event_hld(accel_dev, ADF_EVENT_START)) {
> @@ -267,6 +274,7 @@ static void adf_dev_stop(struct adf_accel_dev *accel_dev)
> !test_bit(ADF_STATUS_STARTING, &accel_dev->status))
> return;
>
> + adf_rl_stop(accel_dev);
> adf_dbgfs_rm(accel_dev);
>
> clear_bit(ADF_STATUS_STARTING, &accel_dev->status);
> @@ -353,6 +361,7 @@ static void adf_dev_shutdown(struct adf_accel_dev *accel_dev)
> clear_bit(accel_dev->accel_id, service->init_status);
> }
>
> + adf_rl_exit(accel_dev);
> adf_heartbeat_shutdown(accel_dev);
>
> hw_data->disable_iov(accel_dev);
> diff --git a/drivers/crypto/intel/qat/qat_common/adf_rl.c b/drivers/crypto/intel/qat/qat_common/adf_rl.c
> new file mode 100644
> index 000000000000..dec98a056dce
> --- /dev/null
> +++ b/drivers/crypto/intel/qat/qat_common/adf_rl.c
> @@ -0,0 +1,1176 @@
> +// SPDX-License-Identifier: GPL-2.0-only
> +/* Copyright(c) 2023 Intel Corporation */
> +
> +#define dev_fmt(fmt) "RateLimiting: " fmt
> +
> +#include <asm/errno.h>
> +#include <asm/div64.h>
> +
> +#include <linux/dev_printk.h>
> +#include <linux/kernel.h>
> +#include <linux/pci.h>
> +#include <linux/slab.h>
> +#include <linux/units.h>
> +
> +#include "adf_accel_devices.h"
> +#include "adf_common_drv.h"
> +#include "adf_rl_admin.h"
> +#include "adf_rl.h"
> +
> +#define RL_TOKEN_GRANULARITY_PCIEIN_BUCKET 0U
> +#define RL_TOKEN_GRANULARITY_PCIEOUT_BUCKET 0U
> +#define RL_TOKEN_PCIE_SIZE 64
> +#define RL_TOKEN_ASYM_SIZE 1024
> +#define RL_CSR_SIZE 4U
> +#define RL_CAPABILITY_MASK GENMASK(6, 4)
> +#define RL_CAPABILITY_VALUE 0x70
> +#define RL_VALIDATE_NON_ZERO(input) ((input) == 0)
> +#define ROOT_MASK GENMASK(1, 0)
> +#define CLUSTER_MASK GENMASK(3, 0)
> +#define LEAF_MASK GENMASK(5, 0)
> +
> +static int validate_user_input(struct adf_accel_dev *accel_dev,
> + struct adf_rl_sla_input_data *sla_in,
> + bool is_update)
> +{
> + const unsigned long rp_mask = sla_in->rp_mask;
> + size_t rp_mask_size;
> + int i, cnt;
> +
> + if (sla_in->sla_id < RL_SLA_EMPTY_ID || sla_in->sla_id >= RL_NODES_CNT_MAX) {

Should this be <= RL_SLA_EMPTY_ID? Additionally this check seems
unnecessary as similar check is done in the validate_sla_id() which gets
always called in the same path where the validate_user_input() is called.

> + dev_notice(&GET_DEV(accel_dev), "Wrong SLA ID\n");
> + goto ret_inval;
> + }
> +
> + if (sla_in->pir < sla_in->cir) {
> + dev_notice(&GET_DEV(accel_dev),
> + "PIR must be >= CIR, setting PIR to CIR\n");
> + sla_in->pir = sla_in->cir;
> + }
> +
> + if (!is_update) {
> + cnt = 0;
> + rp_mask_size = sizeof(sla_in->rp_mask) * BITS_PER_BYTE;
> + for_each_set_bit(i, &rp_mask, rp_mask_size) {
> + if (++cnt > RL_RP_CNT_PER_LEAF_MAX) {
> + dev_notice(&GET_DEV(accel_dev),
> + "Too many ring pairs selected for this SLA\n");
> + goto ret_inval;

The error gotos seem useless in this function, maybe just directly
return -EINVAL.

> + }
> + }
> +
> + if (sla_in->srv >= ADF_SVC_NONE) {
> + dev_notice(&GET_DEV(accel_dev),
> + "Wrong service type\n");
> + goto ret_inval;
> + }
> +
> + if (sla_in->type > RL_LEAF) {
> + dev_notice(&GET_DEV(accel_dev),
> + "Wrong node type\n");
> + goto ret_inval;
> + }
> +
> + if (sla_in->parent_id < RL_PARENT_DEFAULT_ID ||
> + sla_in->parent_id >= RL_NODES_CNT_MAX) {
> + dev_notice(&GET_DEV(accel_dev),
> + "Wrong parent ID\n");
> + goto ret_inval;
> + }
> + }
> +
> + return 0;
> +
> +ret_inval:
> + return -EINVAL;
> +}
> +
> +static int validate_sla_id(struct adf_accel_dev *accel_dev, int sla_id)
> +{
> + struct rl_sla *sla;
> +
> + if (sla_id < 0 || sla_id >= RL_NODES_CNT_MAX)
> + goto ret_not_exists;
> +
> + sla = accel_dev->rate_limiting->sla[sla_id];
> +
> + if (!sla)
> + goto ret_not_exists;
> +
> + if (sla->type != RL_LEAF) {
> + dev_notice(&GET_DEV(accel_dev),
> + "This ID is reserved for internal use\n");
> + goto ret_inval;

Maybe just direct return here.

> + }
> +
> + return 0;
> +
> +ret_not_exists:
> + dev_notice(&GET_DEV(accel_dev),
> + "SLA with provided ID does not exist\n");
> +ret_inval:
> + return -EINVAL;
> +}
> +
> +/**
> + * find_parent() - Find the parent for a new SLA
> + * @rl_data: pointer to ratelimiting data
> + * @sla_in: pointer to user input data for a new SLA
> + *
> + * Function returns a pointer to the parent SLA. If the parent ID is provided
> + * as input in the user data, then such ID is validated and the parent SLA
> + * is returned.
> + * Otherwise, it returns the default parent SLA (root or cluster) for
> + * the new object.
> + *
> + * Return:
> + * * Pointer to the parent SLA object
> + * * NULL - when parent cannot be found
> + */
> +static struct rl_sla *find_parent(struct adf_rl *rl_data,
> + struct adf_rl_sla_input_data *sla_in)
> +{
> + int input_parent_id = sla_in->parent_id;
> + struct rl_sla *root = NULL;
> + struct rl_sla *parent_sla;
> + int i;
> +
> + if (sla_in->type == RL_ROOT)
> + return NULL;
> +
> + if (input_parent_id > RL_PARENT_DEFAULT_ID) {
> + parent_sla = rl_data->sla[input_parent_id];
> + /*
> + * SLA can be a parent if it has the same service as the child
> + * and its type is higher in the hierarchy,
> + * for example the parent type of a LEAF must be a CLUSTER.
> + */
> + if (parent_sla && parent_sla->srv == sla_in->srv &&
> + parent_sla->type == sla_in->type - 1)
> + return parent_sla;
> +
> + return NULL;
> + }
> +
> + /* If input_parent_id is not valid, get root for this service type. */
> + for (i = 0; i < RL_ROOT_MAX; i++) {
> + if (rl_data->root[i] && rl_data->root[i]->srv == sla_in->srv) {
> + root = rl_data->root[i];
> + break;
> + }
> + }
> +
> + if (!root)
> + return NULL;
> +
> + /*
> + * If the type of this SLA is cluster, then return the root.
> + * Otherwise, find the default (i.e. first) cluster for this service.
> + */
> + if (sla_in->type == RL_CLUSTER)
> + return root;
> +
> + for (i = 0; i < RL_CLUSTER_MAX; i++) {
> + if (rl_data->cluster[i] && rl_data->cluster[i]->parent == root)
> + return rl_data->cluster[i];
> + }
> +
> + return NULL;
> +}
> +
> +static enum adf_cfg_service_type srv_to_cfg_svc_type(enum adf_base_services rl_srv)
> +{
> + switch (rl_srv) {
> + case ADF_SVC_ASYM:
> + return ASYM;
> + case ADF_SVC_SYM:
> + return SYM;
> + case ADF_SVC_DC:
> + return COMP;
> + default:
> + return UNUSED;
> + }
> +}
> +
> +/**
> + * get_sla_arr_of_type() - Returns a pointer to SLA type specific array
> + * @rl_data: pointer to ratelimiting data
> + * @type: SLA type
> + * @sla_arr: pointer to variable where requested pointer will be stored
> + *
> + * Return: Max number of elements allowed for the returned array
> + */
> +static u32 get_sla_arr_of_type(struct adf_rl *rl_data, enum rl_node_type type,
> + struct rl_sla ***sla_arr)
> +{
> + switch (type) {
> + case RL_LEAF:
> + *sla_arr = rl_data->leaf;
> + return RL_LEAF_MAX;
> + case RL_CLUSTER:
> + *sla_arr = rl_data->cluster;
> + return RL_CLUSTER_MAX;
> + case RL_ROOT:
> + *sla_arr = rl_data->root;
> + return RL_ROOT_MAX;
> + default:
> + *sla_arr = NULL;
> + return 0;
> + }
> +}
> +
> +static bool is_service_enabled(struct adf_accel_dev *accel_dev,
> + enum adf_base_services rl_srv)
> +{
> + enum adf_cfg_service_type arb_srv = srv_to_cfg_svc_type(rl_srv);
> + struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev);
> + u8 rps_per_bundle = hw_data->num_banks_per_vf;
> + int i;
> +
> + for (i = 0; i < rps_per_bundle; i++) {
> + if (GET_SRV_TYPE(accel_dev, i) == arb_srv)
> + return true;
> + }
> +
> + return false;
> +}
> +
> +/**
> + * prepare_rp_ids() - Creates an array of ring pair IDs from bitmask
> + * @accel_dev: pointer to acceleration device structure
> + * @sla: SLA object data where result will be written
> + * @rp_mask: bitmask of ring pair IDs
> + *
> + * Function tries to convert provided bitmap to an array of IDs. It checks if
> + * RPs aren't in use, are assigned to SLA service or if a number of provided
> + * IDs is not too big. If successful, writes the result into the field
> + * sla->ring_pairs_cnt.
> + *
> + * Return:
> + * * 0 - ok
> + * * -EINVAL - ring pairs array cannot be created from provided mask
> + */
> +static int prepare_rp_ids(struct adf_accel_dev *accel_dev, struct rl_sla *sla,
> + const unsigned long rp_mask)
> +{
> + enum adf_cfg_service_type arb_srv = srv_to_cfg_svc_type(sla->srv);
> + u16 rps_per_bundle = GET_HW_DATA(accel_dev)->num_banks_per_vf;
> + bool *rp_in_use = accel_dev->rate_limiting->rp_in_use;
> + size_t rp_cnt_max = ARRAY_SIZE(sla->ring_pairs_ids);
> + u16 rp_id_max = GET_HW_DATA(accel_dev)->num_banks;
> + u16 cnt = 0;
> + u16 rp_id;
> +
> + for_each_set_bit(rp_id, &rp_mask, rp_id_max) {
> + if (cnt >= rp_cnt_max) {
> + dev_notice(&GET_DEV(accel_dev),
> + "Assigned more ring pairs than supported");
> + return -EINVAL;
> + }
> +
> + if (rp_in_use[rp_id]) {
> + dev_notice(&GET_DEV(accel_dev),
> + "RP %u already assigned to other SLA", rp_id);
> + return -EINVAL;
> + }
> +
> + if (GET_SRV_TYPE(accel_dev, rp_id % rps_per_bundle) != arb_srv) {
> + dev_notice(&GET_DEV(accel_dev),
> + "RP %u does not support SLA service", rp_id);
> + return -EINVAL;
> + }
> +
> + sla->ring_pairs_ids[cnt++] = rp_id;
> + }
> +
> + sla->ring_pairs_cnt = cnt;
> +
> + return 0;
> +}
> +
> +static void mark_rps_usage(struct rl_sla *sla, bool *rp_in_use, bool used)
> +{
> + u16 rp_id;
> + int i;
> +
> + for (i = 0; i < sla->ring_pairs_cnt; i++) {
> + rp_id = sla->ring_pairs_ids[i];
> + rp_in_use[rp_id] = used;
> + }
> +}
> +
> +static void assign_rps_to_leaf(struct adf_accel_dev *accel_dev,
> + struct rl_sla *sla, bool clear)
> +{
> + struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev);
> + void __iomem *pmisc_addr = adf_get_pmisc_base(accel_dev);
> + u32 base_offset = hw_data->rl_data.r2l_offset;
> + u32 node_id = clear ? 0U : (sla->node_id & LEAF_MASK);
> + u32 offset;
> + int i;
> +
> + for (i = 0; i < sla->ring_pairs_cnt; i++) {
> + offset = base_offset + (RL_CSR_SIZE * sla->ring_pairs_ids[i]);
> + ADF_CSR_WR(pmisc_addr, offset, node_id);
> + }
> +}
> +
> +static void assign_leaf_to_cluster(struct adf_accel_dev *accel_dev,
> + struct rl_sla *sla, bool clear)
> +{
> + struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev);
> + void __iomem *pmisc_addr = adf_get_pmisc_base(accel_dev);
> + u32 base_offset = hw_data->rl_data.l2c_offset;
> + u32 node_id = sla->node_id & LEAF_MASK;
> + u32 parent_id = clear ? 0U : (sla->parent->node_id & CLUSTER_MASK);
> + u32 offset;
> +
> + offset = base_offset + (RL_CSR_SIZE * node_id);
> + ADF_CSR_WR(pmisc_addr, offset, parent_id);
> +}
> +
> +static void assign_cluster_to_root(struct adf_accel_dev *accel_dev,
> + struct rl_sla *sla, bool clear)
> +{
> + struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev);
> + void __iomem *pmisc_addr = adf_get_pmisc_base(accel_dev);
> + u32 base_offset = hw_data->rl_data.c2s_offset;
> + u32 node_id = sla->node_id & CLUSTER_MASK;
> + u32 parent_id = clear ? 0U : (sla->parent->node_id & ROOT_MASK);
> + u32 offset;
> +
> + offset = base_offset + (RL_CSR_SIZE * node_id);
> + ADF_CSR_WR(pmisc_addr, offset, parent_id);
> +}
> +
> +static void assign_node_to_parent(struct adf_accel_dev *accel_dev,
> + struct rl_sla *sla, bool clear_assignment)
> +{
> + switch (sla->type) {
> + case RL_LEAF:
> + assign_rps_to_leaf(accel_dev, sla, clear_assignment);
> + assign_leaf_to_cluster(accel_dev, sla, clear_assignment);
> + break;
> + case RL_CLUSTER:
> + assign_cluster_to_root(accel_dev, sla, clear_assignment);
> + break;
> + default:
> + break;
> + }
> +}
> +
> +/**
> + * can_parent_afford_sla() - Verifies if parent allows to create an SLA
> + * @sla_in: pointer to user input data for a new SLA
> + * @sla_parent: pointer to parent SLA object
> + * @sla_cir: current child CIR value (only for update)
> + * @is_update: request is a update
> + *
> + * Algorithm verifies if parent has enough remaining budget to take assignment
> + * of a child with provided parameters. In update case current CIR value must be
> + * returned to budget first.
> + * PIR value cannot exceed the PIR assigned to parent.
> + *
> + * Return:
> + * * true - SLA can be created
> + * * false - SLA cannot be created
> + */
> +static bool can_parent_afford_sla(struct adf_rl_sla_input_data *sla_in,
> + struct rl_sla *sla_parent, u32 sla_cir,
> + bool is_update)
> +{
> + u32 rem_cir = sla_parent->rem_cir;
> +
> + if (is_update)
> + rem_cir += sla_cir;
> +
> + if (sla_in->cir > rem_cir || sla_in->pir > sla_parent->pir)
> + return false;
> +
> + return true;
> +}
> +
> +/**
> + * can_node_afford_update() - Verifies if SLA can be updated with input data
> + * @sla_in: pointer to user input data for a new SLA
> + * @sla: pointer to SLA object selected for update
> + *
> + * Algorithm verifies if a new CIR value is big enough to satisfy currently
> + * assigned child SLAs and if PIR can be updated
> + *
> + * Return:
> + * * true - SLA can be updated
> + * * false - SLA cannot be updated
> + */
> +static bool can_node_afford_update(struct adf_rl_sla_input_data *sla_in,
> + struct rl_sla *sla)
> +{
> + u32 cir_in_use = sla->cir - sla->rem_cir;
> +
> + /* new CIR cannot be smaller then currently consumed value */
> + if (cir_in_use > sla_in->cir)
> + return false;
> +
> + /* PIR of root/cluster cannot be reduced in node with assigned children */
> + if (sla_in->pir < sla->pir && sla->type != RL_LEAF && cir_in_use > 0)
> + return false;
> +
> + return true;
> +}
> +
> +static bool is_enough_budget(struct adf_rl *rl_data, struct rl_sla *sla,
> + struct adf_rl_sla_input_data *sla_in,
> + bool is_update)
> +{
> + u32 max_val = rl_data->device_data->scale_ref;
> + struct rl_sla *parent = sla->parent;
> + bool ret = true;
> +
> + if (sla_in->cir > max_val || sla_in->pir > max_val)
> + ret = false;
> +
> + switch (sla->type) {
> + case RL_LEAF:
> + ret &= can_parent_afford_sla(sla_in, parent, sla->cir,
> + is_update);
> + break;
> + case RL_CLUSTER:
> + ret &= can_parent_afford_sla(sla_in, parent, sla->cir,
> + is_update);
> +
> + if (is_update)
> + ret &= can_node_afford_update(sla_in, sla);
> +
> + break;
> + case RL_ROOT:
> + if (is_update)
> + ret &= can_node_afford_update(sla_in, sla);
> +
> + break;
> + default:
> + ret = false;
> + break;
> + }
> +
> + return ret;
> +}
> +
> +static void update_budget(struct rl_sla *sla, u32 old_cir, bool is_update)
> +{
> + u32 new_rem;
> +
> + switch (sla->type) {
> + case RL_LEAF:
> + if (is_update)
> + sla->parent->rem_cir += old_cir;
> +
> + sla->parent->rem_cir -= sla->cir;
> + sla->rem_cir = 0;
> + break;
> + case RL_CLUSTER:
> + if (is_update) {
> + sla->parent->rem_cir += old_cir;
> + new_rem = sla->cir - (old_cir - sla->rem_cir);
> + sla->rem_cir = new_rem;

Get rid of new_rem and directly assign.

> + } else {
> + sla->rem_cir = sla->cir;
> + }
> +
> + sla->parent->rem_cir -= sla->cir;
> + break;
> + case RL_ROOT:
> + if (is_update) {
> + new_rem = sla->cir - (old_cir - sla->rem_cir);
> + sla->rem_cir = new_rem;

Same here. You avoid one local variable that way.

> + } else {
> + sla->rem_cir = sla->cir;
> + }
> + break;
> + default:
> + break;
> + }
> +}
> +
> +/**
> + * get_next_free_sla_id() - finds next free ID in the SLA array
> + * @rl_data: Pointer to ratelimiting data structure
> + *
> + * Return:
> + * * 0 : RL_NODES_CNT_MAX - correct ID
> + * * -ENOSPC - all SLA slots are in use
> + */
> +static int get_next_free_sla_id(struct adf_rl *rl_data)
> +{
> + int i = 0;
> +
> + while (i < RL_NODES_CNT_MAX && rl_data->sla[i++])
> + ;
> +
> + if (i == RL_NODES_CNT_MAX)
> + return -ENOSPC;
> +
> + return i - 1;
> +}
> +
> +/**
> + * get_next_free_node_id() - finds next free ID in the array of that node type
> + * @rl_data: Pointer to ratelimiting data structure
> + * @sla: Pointer to SLA object for which the ID is searched
> + *
> + * Return:
> + * * 0 : RL_[NODE_TYPE]_MAX - correct ID
> + * * -ENOSPC - all slots of that type are in use
> + */
> +static int get_next_free_node_id(struct adf_rl *rl_data, struct rl_sla *sla)
> +{
> + struct adf_hw_device_data *hw_device = GET_HW_DATA(rl_data->accel_dev);
> + int max_id, i, step, rp_per_leaf;
> + struct rl_sla **sla_list;
> +
> + rp_per_leaf = hw_device->num_banks / hw_device->num_banks_per_vf;
> +
> + /*
> + * Static nodes mapping:
> + * root0 - cluster[0,4,8,12] - leaf[0-15]
> + * root1 - cluster[1,5,9,13] - leaf[16-31]
> + * root2 - cluster[2,6,10,14] - leaf[32-47]
> + */
> + switch (sla->type) {
> + case RL_LEAF:
> + i = sla->srv * rp_per_leaf;
> + step = 1;
> + max_id = i + rp_per_leaf;
> + sla_list = rl_data->leaf;
> + break;
> + case RL_CLUSTER:
> + i = sla->srv;
> + step = 4;
> + max_id = RL_CLUSTER_MAX;
> + sla_list = rl_data->cluster;
> + break;
> + case RL_ROOT:
> + return sla->srv;
> + default:
> + return -EINVAL;
> + }
> +
> + while (i < max_id && sla_list[i])
> + i += step;
> +
> + if (i >= max_id)
> + return -ENOSPC;
> +
> + return i;
> +}
> +
> +u32 adf_rl_calculate_slice_tokens(struct adf_accel_dev *accel_dev, u32 sla_val,
> + enum adf_base_services svc_type)
> +{
> + struct adf_rl_hw_data *device_data = &accel_dev->hw_device->rl_data;
> + struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev);
> + u64 avail_slice_cycles, allocated_tokens;
> +
> + if (!sla_val)
> + return 0;
> +
> + avail_slice_cycles = hw_data->clock_frequency;
> +
> + switch (svc_type) {
> + case ADF_SVC_ASYM:
> + avail_slice_cycles *= device_data->slices.pke_cnt;
> + break;
> + case ADF_SVC_SYM:
> + avail_slice_cycles *= device_data->slices.cph_cnt;
> + break;
> + case ADF_SVC_DC:
> + avail_slice_cycles *= device_data->slices.dcpr_cnt;
> + break;
> + default:
> + break;
> + }
> +
> + do_div(avail_slice_cycles, device_data->scan_interval);
> + allocated_tokens = avail_slice_cycles * sla_val;
> + do_div(allocated_tokens, device_data->scale_ref);
> +
> + return allocated_tokens;
> +}
> +
> +u32 adf_rl_calculate_ae_cycles(struct adf_accel_dev *accel_dev, u32 sla_val,
> + enum adf_base_services svc_type)
> +{
> + struct adf_rl_hw_data *device_data = &accel_dev->hw_device->rl_data;
> + struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev);
> + u64 allocated_ae_cycles, avail_ae_cycles;
> +
> + if (!sla_val)
> + return 0;
> +
> + avail_ae_cycles = hw_data->clock_frequency;
> + avail_ae_cycles *= hw_data->get_num_aes(hw_data) - 1;
> + do_div(avail_ae_cycles, device_data->scan_interval);
> +
> + sla_val *= device_data->max_tp[svc_type];
> + sla_val /= device_data->scale_ref;
> +
> + allocated_ae_cycles = (sla_val * avail_ae_cycles);
> + do_div(allocated_ae_cycles, device_data->max_tp[svc_type]);
> +
> + return allocated_ae_cycles;
> +}
> +
> +u32 adf_rl_calculate_pci_bw(struct adf_accel_dev *accel_dev, u32 sla_val,
> + enum adf_base_services svc_type, bool is_bw_out)
> +{
> + struct adf_rl_hw_data *device_data = &accel_dev->hw_device->rl_data;
> + u64 sla_to_bytes, allocated_bw, sla_scaled;
> +
> + if (!sla_val)
> + return 0;
> +
> + sla_to_bytes = sla_val;
> + sla_to_bytes *= device_data->max_tp[svc_type];
> + do_div(sla_to_bytes, device_data->scale_ref);
> +
> + sla_to_bytes *= (svc_type == ADF_SVC_ASYM) ? RL_TOKEN_ASYM_SIZE :
> + BYTES_PER_MBIT;
> + if (svc_type == ADF_SVC_DC && is_bw_out)
> + sla_to_bytes *= device_data->slices.dcpr_cnt -
> + device_data->dcpr_correction;
> +
> + sla_scaled = sla_to_bytes * device_data->pcie_scale_mul;
> + do_div(sla_scaled, device_data->pcie_scale_div);
> + allocated_bw = sla_scaled;
> + do_div(allocated_bw, RL_TOKEN_PCIE_SIZE);
> + do_div(allocated_bw, device_data->scan_interval);
> +
> + return allocated_bw;
> +}
> +
> +/**
> + * add_new_sla_entry() - creates a new SLA object and fills it with user data
> + * @accel_dev: pointer to acceleration device structure
> + * @sla_in: pointer to user input data for a new SLA
> + * @sla_out: Pointer to variable that will contain the address of a new
> + * SLA object if the operation succeeds
> + *
> + * Return:
> + * * 0 - ok
> + * * -ENOMEM - memory allocation failed
> + * * -EINVAL - invalid user input
> + * * -ENOSPC - all available SLAs are in use
> + */
> +static int add_new_sla_entry(struct adf_accel_dev *accel_dev,
> + struct adf_rl_sla_input_data *sla_in,
> + struct rl_sla **sla_out)
> +{
> + struct adf_rl *rl_data = accel_dev->rate_limiting;
> + struct rl_sla *sla;
> + int ret = 0;
> +
> + sla = kzalloc(sizeof(*sla), GFP_KERNEL);

Why not use devm_kzalloc()? You have access to the device handle.

> + if (!sla) {
> + ret = -ENOMEM;
> + goto ret_err;
> + }
> + *sla_out = sla;
> +
> + if (!is_service_enabled(accel_dev, sla_in->srv)) {
> + dev_notice(&GET_DEV(accel_dev),
> + "Provided service is not enabled\n");
> + ret = -EINVAL;
> + goto ret_err;
> + }
> +
> + sla->srv = sla_in->srv;
> + sla->type = sla_in->type;
> + ret = get_next_free_node_id(rl_data, sla);
> + if (ret < 0) {
> + dev_notice(&GET_DEV(accel_dev),
> + "Exceeded number of available nodes for that service\n");
> + goto ret_err;
> + }
> + sla->node_id = ret;
> +
> + ret = get_next_free_sla_id(rl_data);
> + if (ret < 0) {
> + dev_notice(&GET_DEV(accel_dev),
> + "Allocated maximum SLAs number\n");
> + goto ret_err;
> + }
> + sla->sla_id = ret;
> +
> + sla->parent = find_parent(rl_data, sla_in);
> + if (!sla->parent && sla->type != RL_ROOT) {
> + if (sla_in->parent_id != RL_PARENT_DEFAULT_ID)
> + dev_notice(&GET_DEV(accel_dev),
> + "Provided parent ID does not exist or cannot be parent for this SLA.");
> + else
> + dev_notice(&GET_DEV(accel_dev),
> + "Unable to find parent node for this service. Is service enabled?");
> + ret = -EINVAL;
> + goto ret_err;
> + }
> +
> + if (sla->type == RL_LEAF) {
> + ret = prepare_rp_ids(accel_dev, sla, sla_in->rp_mask);
> + if (!sla->ring_pairs_cnt || ret) {
> + dev_notice(&GET_DEV(accel_dev),
> + "Unable to find ring pairs to assign to the leaf");
> + if (!ret)
> + ret = -EINVAL;
> +
> + goto ret_err;
> + }
> + }
> +
> + ret = 0;
> +
> +ret_err:
> + /* Allocated sla will be freed at the bottom of calling function */

I would rather recommend you release the sla here if the function fails.

> + return ret;
> +}
> +
> +static int initialize_default_nodes(struct adf_accel_dev *accel_dev)
> +{
> + struct adf_rl *rl_data = accel_dev->rate_limiting;
> + struct adf_rl_hw_data *device_data = rl_data->device_data;
> + struct adf_rl_sla_input_data sla_in = { };
> + int ret = 0;
> + int i;
> +
> + /* Init root for each enabled service */
> + sla_in.type = RL_ROOT;
> + sla_in.parent_id = RL_PARENT_DEFAULT_ID;
> +
> + for (i = 0; i < ADF_SVC_NONE; i++) {
> + if (!is_service_enabled(accel_dev, i))
> + continue;
> +
> + sla_in.cir = device_data->scale_ref;
> + sla_in.pir = sla_in.cir;
> + sla_in.srv = i;
> +
> + ret = adf_rl_add_sla(accel_dev, &sla_in);
> + if (ret)
> + goto err_ret;
> + }
> +
> + /* Init default cluster for each root */
> + sla_in.type = RL_CLUSTER;
> + for (i = 0; i < ADF_SVC_NONE; i++) {
> + if (!rl_data->root[i])
> + continue;
> +
> + sla_in.cir = rl_data->root[i]->cir;
> + sla_in.pir = sla_in.cir;
> + sla_in.srv = rl_data->root[i]->srv;
> +
> + ret = adf_rl_add_sla(accel_dev, &sla_in);
> + if (ret)
> + goto err_ret;
> + }
> +
> + return 0;
> +
> +err_ret:
> + dev_notice(&GET_DEV(accel_dev),
> + "Initialization of default nodes failed\n");

dev_notice() here and you also do dev_err() for the exact same reason in
the callee (adf_rl_start().) You can make this function cleaner by
removing the dev_notice here, and directly returning in error cases
without goto.

> + return ret;
> +}
> +
> +static void clear_sla(struct adf_rl *rl_data, struct rl_sla *sla)
> +{
> + bool *rp_in_use = rl_data->rp_in_use;
> + struct rl_sla **sla_type_arr = NULL;
> + int i, sla_id, node_id;
> + u32 old_cir;
> +
> + sla_id = sla->sla_id;
> + node_id = sla->node_id;
> + old_cir = sla->cir;
> + sla->cir = 0;
> + sla->pir = 0;
> +
> + for (i = 0; i < sla->ring_pairs_cnt; i++)
> + rp_in_use[sla->ring_pairs_ids[i]] = false;
> +
> + update_budget(sla, old_cir, true);
> + get_sla_arr_of_type(rl_data, sla->type, &sla_type_arr);
> + assign_node_to_parent(rl_data->accel_dev, sla, true);
> + adf_rl_send_admin_delete_msg(rl_data->accel_dev, node_id, sla->type);
> + mark_rps_usage(sla, rl_data->rp_in_use, false);
> +
> + kfree(sla);
> + rl_data->sla[sla_id] = NULL;
> + sla_type_arr[node_id] = NULL;
> +}
> +
> +/**
> + * add_update_sla() - handles the creation and the update of an SLA
> + * @accel_dev: pointer to acceleration device structure
> + * @sla_in: pointer to user input data for a new/updated SLA
> + * @is_update: flag to indicate if this is an update or an add operation
> + *
> + * Return:
> + * * 0 - ok
> + * * -ENOMEM - memory allocation failed
> + * * -EINVAL - user input data cannot be used to create SLA
> + * * -ENOSPC - all available SLAs are in use
> + */
> +static int add_update_sla(struct adf_accel_dev *accel_dev,
> + struct adf_rl_sla_input_data *sla_in, bool is_update)
> +{
> + struct adf_rl *rl_data = accel_dev->rate_limiting;
> + struct rl_sla **sla_type_arr = NULL;
> + struct rl_sla *sla = NULL;
> + u32 old_cir = 0;
> + int ret;
> +
> + if (!sla_in) {
> + dev_warn(&GET_DEV(accel_dev),
> + "SLA input data pointer is missing\n");
> + ret = -EFAULT;
> + goto ret_err;
> + }
> +
> + /* Input validation */
> + ret = validate_user_input(accel_dev, sla_in, is_update);
> + if (ret)
> + goto ret_err;
> +
> + mutex_lock(&rl_data->rl_lock);
> +
> + if (is_update) {
> + ret = validate_sla_id(accel_dev, sla_in->sla_id);
> + if (ret)
> + goto ret_err;
> +
> + sla = rl_data->sla[sla_in->sla_id];
> + old_cir = sla->cir;
> + } else {
> + ret = add_new_sla_entry(accel_dev, sla_in, &sla);
> + if (ret)
> + goto ret_err;
> + }
> +
> + if (!is_enough_budget(rl_data, sla, sla_in, is_update)) {
> + dev_notice(&GET_DEV(accel_dev),
> + "Input value exceeds the remaining budget%s\n",
> + is_update ? " or more budget is already in use" : "");
> + ret = -EINVAL;
> + goto ret_err;
> + }
> + sla->cir = sla_in->cir;
> + sla->pir = sla_in->pir;
> +
> + /* Apply SLA */
> + assign_node_to_parent(accel_dev, sla, false);
> + ret = adf_rl_send_admin_add_update_msg(accel_dev, sla, is_update);
> + if (ret) {
> + dev_notice(&GET_DEV(accel_dev),
> + "Failed to apply an SLA\n");
> + goto ret_err;
> + }
> + update_budget(sla, old_cir, is_update);
> +
> + if (!is_update) {
> + mark_rps_usage(sla, rl_data->rp_in_use, true);
> + get_sla_arr_of_type(rl_data, sla->type, &sla_type_arr);
> + sla_type_arr[sla->node_id] = sla;
> + rl_data->sla[sla->sla_id] = sla;
> + }
> +
> + sla_in->sla_id = sla->sla_id;
> + goto ret_ok;
> +
> +ret_err:
> + if (!is_update && sla) {
> + sla_in->sla_id = -1;
> + kfree(sla);
> + }
> +ret_ok:
> + mutex_unlock(&rl_data->rl_lock);
> + return ret;
> +}
> +
> +/**
> + * adf_rl_add_sla() - handles the creation of an SLA
> + * @accel_dev: pointer to acceleration device structure
> + * @sla_in: pointer to user input data required to add an SLA
> + *
> + * Return:
> + * * 0 - ok
> + * * -ENOMEM - memory allocation failed
> + * * -EINVAL - invalid user input
> + * * -ENOSPC - all available SLAs are in use
> + */
> +int adf_rl_add_sla(struct adf_accel_dev *accel_dev,
> + struct adf_rl_sla_input_data *sla_in)
> +{
> + return add_update_sla(accel_dev, sla_in, false);
> +}
> +
> +/**
> + * adf_rl_update_sla() - handles the update of an SLA
> + * @accel_dev: pointer to acceleration device structure
> + * @sla_in: pointer to user input data required to update an SLA
> + *
> + * Return:
> + * * 0 - ok
> + * * -EINVAL - user input data cannot be used to update SLA
> + */
> +int adf_rl_update_sla(struct adf_accel_dev *accel_dev,
> + struct adf_rl_sla_input_data *sla_in)
> +{
> + return add_update_sla(accel_dev, sla_in, true);
> +}
> +
> +/**
> + * adf_rl_get_sla() - returns an existing SLA data
> + * @accel_dev: pointer to acceleration device structure
> + * @sla_in: pointer to user data where SLA info will be stored
> + *
> + * The sla_id for which data are requested should be set in sla_id structure
> + *
> + * Return:
> + * * 0 - ok
> + * * -EINVAL - provided sla_id does not exist
> + */
> +int adf_rl_get_sla(struct adf_accel_dev *accel_dev,
> + struct adf_rl_sla_input_data *sla_in)
> +{
> + struct rl_sla *sla;
> + int ret, i;
> +
> + ret = validate_sla_id(accel_dev, sla_in->sla_id);
> + if (ret)
> + return ret;
> +
> + sla = accel_dev->rate_limiting->sla[sla_in->sla_id];
> + sla_in->type = sla->type;
> + sla_in->srv = sla->srv;
> + sla_in->cir = sla->cir;
> + sla_in->pir = sla->pir;
> + sla_in->rp_mask = 0U;
> + if (sla->parent)
> + sla_in->parent_id = sla->parent->sla_id;
> + else
> + sla_in->parent_id = RL_PARENT_DEFAULT_ID;
> +
> + for (i = 0; i < sla->ring_pairs_cnt; i++)
> + sla_in->rp_mask |= BIT(sla->ring_pairs_ids[i]);
> +
> + return 0;
> +}
> +
> +/**
> + * adf_rl_get_capability_remaining() - returns the remaining SLA value (CIR) for
> + * selected service or provided sla_id
> + * @accel_dev: pointer to acceleration device structure
> + * @srv: service ID for which capability is requested
> + * @sla_id: ID of the cluster or root to which we want assign a new SLA
> + *
> + * Check if the provided SLA id is valid. If it is and the service matches
> + * the requested service and the type is cluster or root, return the remaining
> + * capability.
> + * If the provided ID does not match the service or type, return the remaining
> + * capacity of the default cluster for that service.
> + *
> + * Return:
> + * * Positive value - correct remaining value
> + * * -EINVAL - algorithm cannot find a remaining value for provided data
> + */
> +int adf_rl_get_capability_remaining(struct adf_accel_dev *accel_dev,
> + enum adf_base_services srv, int sla_id)
> +{
> + struct adf_rl *rl_data = accel_dev->rate_limiting;
> + struct rl_sla *sla = NULL;
> + int i;
> +
> + if (srv >= ADF_SVC_NONE)
> + return -EINVAL;
> +
> + if (sla_id > RL_SLA_EMPTY_ID && !validate_sla_id(accel_dev, sla_id)) {
> + sla = rl_data->sla[sla_id];
> +
> + if (sla->srv == srv && sla->type <= RL_CLUSTER)
> + goto ret_ok;
> + }
> +
> + for (i = 0; i < RL_CLUSTER_MAX; i++) {
> + if (!rl_data->cluster[i])
> + continue;
> +
> + if (rl_data->cluster[i]->srv == srv) {
> + sla = rl_data->cluster[i];
> + goto ret_ok;
> + }
> + }
> +
> + return -EINVAL;
> +ret_ok:
> + return sla->rem_cir;
> +}
> +
> +/**
> + * adf_rl_remove_sla() - removes provided sla_id
> + * @accel_dev: pointer to acceleration device structure
> + * @sla_id: ID of the cluster or root to which we want assign an new SLA
> + *
> + * Return:
> + * * 0 - ok
> + * * -EINVAL - wrong sla_id or it still have assigned children
> + */
> +int adf_rl_remove_sla(struct adf_accel_dev *accel_dev, u32 sla_id)
> +{
> + struct adf_rl *rl_data = accel_dev->rate_limiting;
> + struct rl_sla *sla;
> + int ret;
> +
> + ret = validate_sla_id(accel_dev, sla_id);
> + if (ret)
> + return ret;
> +
> + sla = rl_data->sla[sla_id];
> +
> + if (sla->type < RL_LEAF && sla->rem_cir != sla->cir) {
> + dev_notice(&GET_DEV(accel_dev),
> + "To remove parent SLA all its children must be removed first");
> + return -EINVAL;
> + }

Mutex is not protecting the above portion of code, which means you can
call clear_sla() with identical pointer value multiple times.

> +
> + mutex_lock(&rl_data->rl_lock);
> + clear_sla(rl_data, sla);
> + mutex_unlock(&rl_data->rl_lock);
> +
> + return 0;
> +}
> +
> +/**
> + * adf_rl_remove_sla_all() - removes all SLAs from device
> + * @accel_dev: pointer to acceleration device structure
> + * @incl_default: set to true if default SLAs also should be removed
> + */
> +void adf_rl_remove_sla_all(struct adf_accel_dev *accel_dev, bool incl_default)
> +{
> + struct adf_rl *rl_data = accel_dev->rate_limiting;
> + int end_type = incl_default ? RL_ROOT : RL_LEAF;
> + struct rl_sla **sla_type_arr = NULL;
> + u32 max_id;
> + int i, j;
> +
> + mutex_lock(&rl_data->rl_lock);
> +
> + /* Unregister and remove all SLAs */
> + for (j = RL_LEAF; j >= end_type; j--) {
> + max_id = get_sla_arr_of_type(rl_data, j, &sla_type_arr);
> +
> + for (i = 0; i < max_id; i++) {
> + if (!sla_type_arr[i])
> + continue;
> +
> + clear_sla(rl_data, sla_type_arr[i]);
> + }
> + }
> +
> + mutex_unlock(&rl_data->rl_lock);
> +}
> +
> +int adf_rl_init(struct adf_accel_dev *accel_dev)
> +{
> + struct adf_hw_device_data *hw_data = GET_HW_DATA(accel_dev);
> + struct adf_rl_hw_data *rl_hw_data = &hw_data->rl_data;
> + struct adf_rl *rl;
> + int ret = 0;
> +
> + /* Validate device parameters */
> + if (RL_VALIDATE_NON_ZERO(rl_hw_data->max_tp[ADF_SVC_ASYM]) ||
> + RL_VALIDATE_NON_ZERO(rl_hw_data->max_tp[ADF_SVC_SYM]) ||
> + RL_VALIDATE_NON_ZERO(rl_hw_data->max_tp[ADF_SVC_DC]) ||
> + RL_VALIDATE_NON_ZERO(rl_hw_data->scan_interval) ||
> + RL_VALIDATE_NON_ZERO(rl_hw_data->pcie_scale_div) ||
> + RL_VALIDATE_NON_ZERO(rl_hw_data->pcie_scale_mul) ||
> + RL_VALIDATE_NON_ZERO(rl_hw_data->scale_ref)) {
> + ret = -EOPNOTSUPP;
> + goto err_ret;
> + }
> +
> + rl = kzalloc(sizeof(*rl), GFP_KERNEL);

devm_kzalloc()?

> + if (!rl) {
> + ret = -ENOMEM;
> + goto err_ret;
> + }
> +
> + mutex_init(&rl->rl_lock);
> + rl->device_data = &accel_dev->hw_device->rl_data;
> + rl->accel_dev = accel_dev;
> + accel_dev->rate_limiting = rl;
> +
> +err_ret:
> + return ret;
> +}
> +
> +int adf_rl_start(struct adf_accel_dev *accel_dev)
> +{
> + struct adf_rl_hw_data *rl_hw_data = &GET_HW_DATA(accel_dev)->rl_data;
> + void __iomem *pmisc_addr = adf_get_pmisc_base(accel_dev);
> + u16 fw_caps = GET_HW_DATA(accel_dev)->fw_capabilities;
> + int ret;
> +
> + if (!accel_dev->rate_limiting) {
> + ret = -EOPNOTSUPP;
> + goto ret_err;
> + }
> +
> + if ((fw_caps & RL_CAPABILITY_MASK) != RL_CAPABILITY_VALUE) {
> + dev_info(&GET_DEV(accel_dev), "not supported\n");
> + ret = -EOPNOTSUPP;
> + goto ret_free;
> + }
> +
> + ADF_CSR_WR(pmisc_addr, rl_hw_data->pciin_tb_offset,
> + RL_TOKEN_GRANULARITY_PCIEIN_BUCKET);
> + ADF_CSR_WR(pmisc_addr, rl_hw_data->pciout_tb_offset,
> + RL_TOKEN_GRANULARITY_PCIEOUT_BUCKET);
> +
> + ret = adf_rl_send_admin_init_msg(accel_dev, &rl_hw_data->slices);
> + if (ret) {
> + dev_err(&GET_DEV(accel_dev), "initialization failed\n");
> + goto ret_free;
> + }
> +
> + ret = initialize_default_nodes(accel_dev);
> + if (ret) {
> + dev_err(&GET_DEV(accel_dev),
> + "failed to initialize default SLAs\n");
> + goto ret_sla_rm;
> + }
> +
> + return 0;
> +
> +ret_sla_rm:
> + adf_rl_remove_sla_all(accel_dev, true);
> +ret_free:
> + kfree(accel_dev->rate_limiting);
> + accel_dev->rate_limiting = NULL;
> +ret_err:
> + return ret;
> +}
> +
> +void adf_rl_stop(struct adf_accel_dev *accel_dev)
> +{
> + if (!accel_dev->rate_limiting)
> + return;
> +
> + adf_rl_remove_sla_all(accel_dev, true);
> +}
> +
> +void adf_rl_exit(struct adf_accel_dev *accel_dev)
> +{
> + if (!accel_dev->rate_limiting)
> + return;
> +
> + kfree(accel_dev->rate_limiting);
> + accel_dev->rate_limiting = NULL;
> +}
> diff --git a/drivers/crypto/intel/qat/qat_common/adf_rl.h b/drivers/crypto/intel/qat/qat_common/adf_rl.h
> new file mode 100644
> index 000000000000..1ccb6613c92e
> --- /dev/null
> +++ b/drivers/crypto/intel/qat/qat_common/adf_rl.h
> @@ -0,0 +1,169 @@
> +/* SPDX-License-Identifier: GPL-2.0-only */
> +/* Copyright(c) 2023 Intel Corporation */
> +
> +#ifndef ADF_RL_H_
> +#define ADF_RL_H_
> +
> +#include <linux/mutex.h>
> +#include <linux/types.h>
> +
> +struct adf_accel_dev;
> +
> +#define RL_ROOT_MAX 4
> +#define RL_CLUSTER_MAX 16
> +#define RL_LEAF_MAX 64
> +#define RL_NODES_CNT_MAX (RL_ROOT_MAX + RL_CLUSTER_MAX + RL_LEAF_MAX)
> +#define RL_RP_CNT_PER_LEAF_MAX 4U
> +#define RL_RP_CNT_MAX 64
> +#define RL_SLA_EMPTY_ID -1
> +#define RL_PARENT_DEFAULT_ID -1
> +
> +enum rl_node_type {
> + RL_ROOT,
> + RL_CLUSTER,
> + RL_LEAF,
> +};
> +
> +enum adf_base_services {
> + ADF_SVC_ASYM = 0,
> + ADF_SVC_SYM,
> + ADF_SVC_DC,
> + ADF_SVC_NONE,
> +};
> +
> +/**
> + * struct adf_rl_sla_input_data - ratelimiting user input data structure
> + * @rp_mask: 64 bit bitmask of ring pair IDs which will be assigned to SLA.
> + * Eg. 0x5 -> RP0 and RP2 assigned; 0xA005 -> RP0,2,13,15 assigned.
> + * @sla_id: ID of current SLA for operations update, rm, get. For the add
> + * operation, this field will be updated with the ID of the newly
> + * added SLA
> + * @parent_id: ID of the SLA to which the current one should be assigned.
> + * Set to -1 to refer to the default parent.
> + * @cir: Committed information rate. Rate guaranteed to be achieved. Input value
> + * is expressed in permille scale, i.e. 1000 refers to the maximum
> + * device throughput for a selected service.
> + * @pir: Peak information rate. Maximum rate available that the SLA can achieve.
> + * Input value is expressed in permille scale, i.e. 1000 refers to
> + * the maximum device throughput for a selected service.
> + * @type: SLA type: root, cluster, node
> + * @srv: Service associated to the SLA: asym, sym dc.
> + *
> + * This structure is used to perform operations on an SLA.
> + * Depending on the operation, some of the parameters are ignored.
> + * The following list reports which parameters should be set for each operation.
> + * - add: all except sla_id
> + * - update: cir, pir, sla_id
> + * - rm: sla_id
> + * - rm_all: -
> + * - get: sla_id
> + * - get_capability_rem: srv, sla_id
> + */
> +struct adf_rl_sla_input_data {
> + u64 rp_mask;
> + int sla_id;
> + int parent_id;
> + unsigned int cir;
> + unsigned int pir;
> + enum rl_node_type type;
> + enum adf_base_services srv;
> +};
> +
> +struct rl_slice_cnt {
> + u8 dcpr_cnt;
> + u8 pke_cnt;
> + u8 cph_cnt;
> +};
> +
> +struct adf_rl_hw_data {
> + u32 scale_ref;
> + u32 scan_interval;
> + u32 r2l_offset;
> + u32 l2c_offset;
> + u32 c2s_offset;
> + u32 pciin_tb_offset;
> + u32 pciout_tb_offset;
> + u32 pcie_scale_mul;
> + u32 pcie_scale_div;
> + u32 dcpr_correction;
> + u32 max_tp[RL_ROOT_MAX];
> + struct rl_slice_cnt slices;
> +};
> +
> +/**
> + * struct adf_rl - ratelimiting data structure
> + * @accel_dev: pointer to acceleration device data
> + * @device_data: pointer to rate limiting data specific to a device type (or revision)
> + * @sla: array of pointers to SLA objects
> + * @root: array of pointers to root type SLAs, element number reflects node_id
> + * @cluster: array of pointers to cluster type SLAs, element number reflects node_id
> + * @leaf: array of pointers to leaf type SLAs, element number reflects node_id
> + * @rp_in_use: array of ring pair IDs already used in one of SLAs
> + * @rl_lock: mutex object which is protecting data in this structure
> + * @input: structure which is used for holding the data received from user
> + */
> +struct adf_rl {
> + struct adf_accel_dev *accel_dev;
> + struct adf_rl_hw_data *device_data;
> + /* mapping sla_id to SLA objects */
> + struct rl_sla *sla[RL_NODES_CNT_MAX];
> + struct rl_sla *root[RL_ROOT_MAX];
> + struct rl_sla *cluster[RL_CLUSTER_MAX];
> + struct rl_sla *leaf[RL_LEAF_MAX];
> + bool rp_in_use[RL_RP_CNT_MAX];
> + /* Mutex protecting writing to SLAs lists */
> + struct mutex rl_lock;
> +};
> +
> +/**
> + * struct rl_sla - SLA object data structure
> + * @parent: pointer to the parent SLA (root/cluster)
> + * @type: SLA type
> + * @srv: service associated with this SLA
> + * @sla_id: ID of the SLA, used as element number in SLA array and as identifier
> + * shared with the user
> + * @node_id: ID of node, each of SLA type have a separate ID list
> + * @cir: committed information rate
> + * @pir: peak information rate (PIR >= CIR)
> + * @rem_cir: if this SLA is a parent then this field represents a remaining
> + * value to be used by child SLAs.
> + * @ring_pairs_ids: array with numeric ring pairs IDs assigned to this SLA
> + * @ring_pairs_cnt: number of assigned ring pairs listed in the array above
> + */
> +struct rl_sla {
> + struct rl_sla *parent;
> + enum rl_node_type type;
> + enum adf_base_services srv;
> + u32 sla_id;
> + u32 node_id;
> + u32 cir;
> + u32 pir;
> + u32 rem_cir;
> + u16 ring_pairs_ids[RL_RP_CNT_PER_LEAF_MAX];
> + u16 ring_pairs_cnt;
> +};
> +
> +int adf_rl_add_sla(struct adf_accel_dev *accel_dev,
> + struct adf_rl_sla_input_data *sla_in);
> +int adf_rl_update_sla(struct adf_accel_dev *accel_dev,
> + struct adf_rl_sla_input_data *sla_in);
> +int adf_rl_get_sla(struct adf_accel_dev *accel_dev,
> + struct adf_rl_sla_input_data *sla_in);
> +int adf_rl_get_capability_remaining(struct adf_accel_dev *accel_dev,
> + enum adf_base_services srv, int sla_id);
> +int adf_rl_remove_sla(struct adf_accel_dev *accel_dev, u32 sla_id);
> +void adf_rl_remove_sla_all(struct adf_accel_dev *accel_dev, bool incl_default);
> +
> +int adf_rl_init(struct adf_accel_dev *accel_dev);
> +int adf_rl_start(struct adf_accel_dev *accel_dev);
> +void adf_rl_stop(struct adf_accel_dev *accel_dev);
> +void adf_rl_exit(struct adf_accel_dev *accel_dev);
> +
> +u32 adf_rl_calculate_pci_bw(struct adf_accel_dev *accel_dev, u32 sla_val,
> + enum adf_base_services svc_type, bool is_bw_out);
> +u32 adf_rl_calculate_ae_cycles(struct adf_accel_dev *accel_dev, u32 sla_val,
> + enum adf_base_services svc_type);
> +u32 adf_rl_calculate_slice_tokens(struct adf_accel_dev *accel_dev, u32 sla_val,
> + enum adf_base_services svc_type);
> +
> +#endif /* ADF_RL_H_ */
> diff --git a/drivers/crypto/intel/qat/qat_common/adf_rl_admin.c b/drivers/crypto/intel/qat/qat_common/adf_rl_admin.c
> new file mode 100644
> index 000000000000..820fcb2d42bf
> --- /dev/null
> +++ b/drivers/crypto/intel/qat/qat_common/adf_rl_admin.c
> @@ -0,0 +1,98 @@
> +// SPDX-License-Identifier: GPL-2.0-only
> +/* Copyright(c) 2023 Intel Corporation */
> +
> +#include <linux/dma-mapping.h>
> +#include <linux/pci.h>
> +
> +#include "adf_admin.h"
> +#include "adf_accel_devices.h"
> +#include "adf_rl_admin.h"
> +
> +static void
> +prep_admin_req_msg(struct rl_sla *sla, dma_addr_t dma_addr,
> + struct icp_qat_fw_init_admin_sla_config_params *fw_params,
> + struct icp_qat_fw_init_admin_req *req, bool is_update)
> +{
> + req->cmd_id = is_update ? ICP_QAT_FW_RL_UPDATE : ICP_QAT_FW_RL_ADD;
> + req->init_cfg_ptr = dma_addr;
> + req->init_cfg_sz = sizeof(*fw_params);
> + req->node_id = sla->node_id;
> + req->node_type = sla->type;
> + req->rp_count = sla->ring_pairs_cnt;
> + req->svc_type = sla->srv;
> +}
> +
> +static void
> +prep_admin_req_params(struct adf_accel_dev *accel_dev, struct rl_sla *sla,
> + struct icp_qat_fw_init_admin_sla_config_params *fw_params)
> +{
> + fw_params->pcie_in_cir =
> + adf_rl_calculate_pci_bw(accel_dev, sla->cir, sla->srv, false);
> + fw_params->pcie_in_pir =
> + adf_rl_calculate_pci_bw(accel_dev, sla->pir, sla->srv, false);
> + fw_params->pcie_out_cir =
> + adf_rl_calculate_pci_bw(accel_dev, sla->cir, sla->srv, true);
> + fw_params->pcie_out_pir =
> + adf_rl_calculate_pci_bw(accel_dev, sla->pir, sla->srv, true);
> +
> + fw_params->slice_util_cir =
> + adf_rl_calculate_slice_tokens(accel_dev, sla->cir, sla->srv);
> + fw_params->slice_util_pir =
> + adf_rl_calculate_slice_tokens(accel_dev, sla->pir, sla->srv);
> +
> + fw_params->ae_util_cir =
> + adf_rl_calculate_ae_cycles(accel_dev, sla->cir, sla->srv);
> + fw_params->ae_util_pir =
> + adf_rl_calculate_ae_cycles(accel_dev, sla->pir, sla->srv);
> +
> + memcpy(fw_params->rp_ids, sla->ring_pairs_ids,
> + sizeof(sla->ring_pairs_ids));
> +}
> +
> +int adf_rl_send_admin_init_msg(struct adf_accel_dev *accel_dev,
> + struct rl_slice_cnt *slices_int)
> +{
> + struct icp_qat_fw_init_admin_slice_cnt slices_resp = { };
> + int ret;
> +
> + ret = adf_send_admin_rl_init(accel_dev, &slices_resp);
> + if (ret) > + goto err_ret;

Just return ret.

-Tero

> +
> + slices_int->dcpr_cnt = slices_resp.dcpr_cnt;
> + slices_int->pke_cnt = slices_resp.pke_cnt;
> + /* For symmetric crypto, slice tokens are relative to the UCS slice */
> + slices_int->cph_cnt = slices_resp.ucs_cnt;
> +
> +err_ret:
> + return ret;
> +}
> +
> +int adf_rl_send_admin_add_update_msg(struct adf_accel_dev *accel_dev,
> + struct rl_sla *sla, bool is_update)
> +{
> + struct icp_qat_fw_init_admin_sla_config_params *fw_params;
> + struct icp_qat_fw_init_admin_req req = { };
> + dma_addr_t dma_addr;
> + int ret;
> +
> + fw_params = dma_alloc_coherent(&GET_DEV(accel_dev), sizeof(*fw_params),
> + &dma_addr, GFP_KERNEL);
> + if (!fw_params)
> + return -ENOMEM;
> +
> + prep_admin_req_params(accel_dev, sla, fw_params);
> + prep_admin_req_msg(sla, dma_addr, fw_params, &req, is_update);
> + ret = adf_send_admin_rl_add_update(accel_dev, &req);
> +
> + dma_free_coherent(&GET_DEV(accel_dev), sizeof(*fw_params), fw_params,
> + dma_addr);
> +
> + return ret;
> +}
> +
> +int adf_rl_send_admin_delete_msg(struct adf_accel_dev *accel_dev, u16 node_id,
> + u8 node_type)
> +{
> + return adf_send_admin_rl_delete(accel_dev, node_id, node_type);
> +}
> diff --git a/drivers/crypto/intel/qat/qat_common/adf_rl_admin.h b/drivers/crypto/intel/qat/qat_common/adf_rl_admin.h
> new file mode 100644
> index 000000000000..dd5419b7e896
> --- /dev/null
> +++ b/drivers/crypto/intel/qat/qat_common/adf_rl_admin.h
> @@ -0,0 +1,18 @@
> +/* SPDX-License-Identifier: GPL-2.0-only */
> +/* Copyright(c) 2023 Intel Corporation */
> +
> +#ifndef ADF_RL_ADMIN_H_
> +#define ADF_RL_ADMIN_H_
> +
> +#include <linux/types.h>
> +
> +#include "adf_rl.h"
> +
> +int adf_rl_send_admin_init_msg(struct adf_accel_dev *accel_dev,
> + struct rl_slice_cnt *slices_int);
> +int adf_rl_send_admin_add_update_msg(struct adf_accel_dev *accel_dev,
> + struct rl_sla *sla, bool is_update);
> +int adf_rl_send_admin_delete_msg(struct adf_accel_dev *accel_dev, u16 node_id,
> + u8 node_type);
> +
> +#endif /* ADF_RL_ADMIN_H_ */
> diff --git a/drivers/crypto/intel/qat/qat_common/icp_qat_fw_init_admin.h b/drivers/crypto/intel/qat/qat_common/icp_qat_fw_init_admin.h
> index e4de9a30e0bd..cd418b51d9f3 100644
> --- a/drivers/crypto/intel/qat/qat_common/icp_qat_fw_init_admin.h
> +++ b/drivers/crypto/intel/qat/qat_common/icp_qat_fw_init_admin.h
> @@ -5,6 +5,8 @@
>
> #include "icp_qat_fw.h"
>
> +#define RL_MAX_RP_IDS 16
> +
> enum icp_qat_fw_init_admin_cmd_id {
> ICP_QAT_FW_INIT_AE = 0,
> ICP_QAT_FW_TRNG_ENABLE = 1,
> @@ -19,10 +21,14 @@ enum icp_qat_fw_init_admin_cmd_id {
> ICP_QAT_FW_CRYPTO_CAPABILITY_GET = 10,
> ICP_QAT_FW_DC_CHAIN_INIT = 11,
> ICP_QAT_FW_HEARTBEAT_TIMER_SET = 13,
> + ICP_QAT_FW_RL_INIT = 15,
> ICP_QAT_FW_TIMER_GET = 19,
> ICP_QAT_FW_CNV_STATS_GET = 20,
> ICP_QAT_FW_PM_STATE_CONFIG = 128,
> ICP_QAT_FW_PM_INFO = 129,
> + ICP_QAT_FW_RL_ADD = 134,
> + ICP_QAT_FW_RL_UPDATE = 135,
> + ICP_QAT_FW_RL_REMOVE = 136,
> };
>
> enum icp_qat_fw_init_admin_resp_status {
> @@ -30,6 +36,30 @@ enum icp_qat_fw_init_admin_resp_status {
> ICP_QAT_FW_INIT_RESP_STATUS_FAIL
> };
>
> +struct icp_qat_fw_init_admin_slice_cnt {
> + __u8 cpr_cnt;
> + __u8 xlt_cnt;
> + __u8 dcpr_cnt;
> + __u8 pke_cnt;
> + __u8 wat_cnt;
> + __u8 wcp_cnt;
> + __u8 ucs_cnt;
> + __u8 cph_cnt;
> + __u8 ath_cnt;
> +};
> +
> +struct icp_qat_fw_init_admin_sla_config_params {
> + __u32 pcie_in_cir;
> + __u32 pcie_in_pir;
> + __u32 pcie_out_cir;
> + __u32 pcie_out_pir;
> + __u32 slice_util_cir;
> + __u32 slice_util_pir;
> + __u32 ae_util_cir;
> + __u32 ae_util_pir;
> + __u16 rp_ids[RL_MAX_RP_IDS];
> +};
> +
> struct icp_qat_fw_init_admin_req {
> __u16 init_cfg_sz;
> __u8 resrvd1;
> @@ -49,6 +79,13 @@ struct icp_qat_fw_init_admin_req {
> struct {
> __u32 heartbeat_ticks;
> };
> + struct {
> + __u16 node_id;
> + __u8 node_type;
> + __u8 svc_type;
> + __u8 resrvd5[3];
> + __u8 rp_count;
> + };
> __u32 idle_filter;
> };
>
> @@ -110,6 +147,7 @@ struct icp_qat_fw_init_admin_resp {
> __u32 unsuccessful_count;
> __u64 resrvd8;
> };
> + struct icp_qat_fw_init_admin_slice_cnt slices;
> __u16 fw_capabilities;
> };
> } __packed;
---------------------------------------------------------------------
Intel Finland Oy
Registered Address: PL 281, 00181 Helsinki
Business Identity Code: 0357606 - 4
Domiciled in Helsinki

This e-mail and any attachments may contain confidential material for
the sole use of the intended recipient(s). Any review or distribution
by others is strictly prohibited. If you are not the intended
recipient, please contact the sender and delete all copies.

2023-10-16 16:53:26

by Damian Muszynski

[permalink] [raw]
Subject: Re: [PATCH 08/11] crypto: qat - add rate limiting feature to qat_4xxx

On 2023-10-16 at 13:53:14 +0300, Tero Kristo wrote:
> Hello,
>
> On 11/10/2023 15:15, Muszynski, Damian wrote:
> > The Rate Limiting (RL) feature allows to control the rate of requests
> > that can be submitted on a ring pair (RP). This allows sharing a QAT
> > device among multiple users while ensuring a guaranteed throughput.

...

> > +static int validate_user_input(struct adf_accel_dev *accel_dev,
> > + struct adf_rl_sla_input_data *sla_in,
> > + bool is_update)
> > +{
> > + const unsigned long rp_mask = sla_in->rp_mask;
> > + size_t rp_mask_size;
> > + int i, cnt;
> > +
> > + if (sla_in->sla_id < RL_SLA_EMPTY_ID || sla_in->sla_id >= RL_NODES_CNT_MAX) {
>
> Should this be <= RL_SLA_EMPTY_ID? Additionally this check seems unnecessary
> as similar check is done in the validate_sla_id() which gets always called
> in the same path where the validate_user_input() is called.

You're right this check can be safely eliminated.

>
> > + dev_notice(&GET_DEV(accel_dev), "Wrong SLA ID\n");
> > + goto ret_inval;
> > + }
> > +
> > + if (sla_in->pir < sla_in->cir) {
> > + dev_notice(&GET_DEV(accel_dev),
> > + "PIR must be >= CIR, setting PIR to CIR\n");
> > + sla_in->pir = sla_in->cir;
> > + }
> > +
> > + if (!is_update) {
> > + cnt = 0;
> > + rp_mask_size = sizeof(sla_in->rp_mask) * BITS_PER_BYTE;
> > + for_each_set_bit(i, &rp_mask, rp_mask_size) {
> > + if (++cnt > RL_RP_CNT_PER_LEAF_MAX) {
> > + dev_notice(&GET_DEV(accel_dev),
> > + "Too many ring pairs selected for this SLA\n");
> > + goto ret_inval;
>
> The error gotos seem useless in this function, maybe just directly return
> -EINVAL.

Ok, will change that.

>
> > + }
> > + }
> > +
> > + if (sla_in->srv >= ADF_SVC_NONE) {
> > + dev_notice(&GET_DEV(accel_dev),
> > + "Wrong service type\n");
> > + goto ret_inval;
> > + }
> > +
> > + if (sla_in->type > RL_LEAF) {
> > + dev_notice(&GET_DEV(accel_dev),
> > + "Wrong node type\n");
> > + goto ret_inval;
> > + }
> > +
> > + if (sla_in->parent_id < RL_PARENT_DEFAULT_ID ||
> > + sla_in->parent_id >= RL_NODES_CNT_MAX) {
> > + dev_notice(&GET_DEV(accel_dev),
> > + "Wrong parent ID\n");
> > + goto ret_inval;
> > + }
> > + }
> > +
> > + return 0;
> > +
> > +ret_inval:
> > + return -EINVAL;
> > +}
> > +
> > +static int validate_sla_id(struct adf_accel_dev *accel_dev, int sla_id)
> > +{
> > + struct rl_sla *sla;
> > +
> > + if (sla_id < 0 || sla_id >= RL_NODES_CNT_MAX)
> > + goto ret_not_exists;
> > +
> > + sla = accel_dev->rate_limiting->sla[sla_id];
> > +
> > + if (!sla)
> > + goto ret_not_exists;
> > +
> > + if (sla->type != RL_LEAF) {
> > + dev_notice(&GET_DEV(accel_dev),
> > + "This ID is reserved for internal use\n");
> > + goto ret_inval;
>
> Maybe just direct return here.

Done.

...

> > +static void update_budget(struct rl_sla *sla, u32 old_cir, bool is_update)
> > +{
> > + u32 new_rem;
> > +
> > + switch (sla->type) {
> > + case RL_LEAF:
> > + if (is_update)
> > + sla->parent->rem_cir += old_cir;
> > +
> > + sla->parent->rem_cir -= sla->cir;
> > + sla->rem_cir = 0;
> > + break;
> > + case RL_CLUSTER:
> > + if (is_update) {
> > + sla->parent->rem_cir += old_cir;
> > + new_rem = sla->cir - (old_cir - sla->rem_cir);
> > + sla->rem_cir = new_rem;
>
> Get rid of new_rem and directly assign.

Ok.

>
> > + } else {
> > + sla->rem_cir = sla->cir;
> > + }
> > +
> > + sla->parent->rem_cir -= sla->cir;
> > + break;
> > + case RL_ROOT:
> > + if (is_update) {
> > + new_rem = sla->cir - (old_cir - sla->rem_cir);
> > + sla->rem_cir = new_rem;
>
> Same here. You avoid one local variable that way.

Ok.

...

> > +static int add_new_sla_entry(struct adf_accel_dev *accel_dev,
> > + struct adf_rl_sla_input_data *sla_in,
> > + struct rl_sla **sla_out)
> > +{
> > + struct adf_rl *rl_data = accel_dev->rate_limiting;
> > + struct rl_sla *sla;
> > + int ret = 0;
> > +
> > + sla = kzalloc(sizeof(*sla), GFP_KERNEL);
>
> Why not use devm_kzalloc()? You have access to the device handle.

Our driver allows user to put a device 'down' and 'up'. During that procedure
many components needs to be unloaded and frees it's memory. This component also
belongs to that group. So devm_* is pointless here.

...

> > + if (sla->type == RL_LEAF) {
> > + ret = prepare_rp_ids(accel_dev, sla, sla_in->rp_mask);
> > + if (!sla->ring_pairs_cnt || ret) {
> > + dev_notice(&GET_DEV(accel_dev),
> > + "Unable to find ring pairs to assign to the leaf");
> > + if (!ret)
> > + ret = -EINVAL;
> > +
> > + goto ret_err;
> > + }
> > + }
> > +
> > + ret = 0;
> > +
> > +ret_err:
> > + /* Allocated sla will be freed at the bottom of calling function */
>
> I would rather recommend you release the sla here if the function fails.

Ok, will add.

...

> > +
> > + ret = adf_rl_add_sla(accel_dev, &sla_in);
> > + if (ret)
> > + goto err_ret;
> > + }
> > +
> > + return 0;
> > +
> > +err_ret:
> > + dev_notice(&GET_DEV(accel_dev),
> > + "Initialization of default nodes failed\n");
>
> dev_notice() here and you also do dev_err() for the exact same reason in the
> callee (adf_rl_start().) You can make this function cleaner by removing the
> dev_notice here, and directly returning in error cases without goto.

Thanks. Will fix that.

...

> > +int adf_rl_remove_sla(struct adf_accel_dev *accel_dev, u32 sla_id)
> > +{
> > + struct adf_rl *rl_data = accel_dev->rate_limiting;
> > + struct rl_sla *sla;
> > + int ret;
> > +
> > + ret = validate_sla_id(accel_dev, sla_id);
> > + if (ret)
> > + return ret;
> > +
> > + sla = rl_data->sla[sla_id];
> > +
> > + if (sla->type < RL_LEAF && sla->rem_cir != sla->cir) {
> > + dev_notice(&GET_DEV(accel_dev),
> > + "To remove parent SLA all its children must be removed first");
> > + return -EINVAL;
> > + }
>
> Mutex is not protecting the above portion of code, which means you can call
> clear_sla() with identical pointer value multiple times.

Thanks. Will fix that.

> > +
> > + mutex_lock(&rl_data->rl_lock);
> > + clear_sla(rl_data, sla);
> > + mutex_unlock(&rl_data->rl_lock);
> > +
> > + return 0;
> > +}

...

> > + rl = kzalloc(sizeof(*rl), GFP_KERNEL);
>
> devm_kzalloc()?

No, explanation above.

...


> > +int adf_rl_send_admin_init_msg(struct adf_accel_dev *accel_dev,
> > + struct rl_slice_cnt *slices_int)
> > +{
> > + struct icp_qat_fw_init_admin_slice_cnt slices_resp = { };
> > + int ret;
> > +
> > + ret = adf_send_admin_rl_init(accel_dev, &slices_resp);
> > + if (ret) > + goto err_ret;
>
> Just return ret.

Ok.


---
Damian Muszynski

2023-10-20 05:27:10

by Herbert Xu

[permalink] [raw]
Subject: Re: [PATCH 00/11] crypto: qat - add rate limiting feature to qat_4xxx

On Wed, Oct 11, 2023 at 02:14:58PM +0200, Damian Muszynski wrote:
> This set introduces the rate limiting feature to the Intel QAT accelerator.
> The Rate Limiting allows control of the rate of the requests that can be
> submitted on a ring pair (RP). This allows sharing a QAT device among
> multiple users while ensuring a guaranteed throughput.

This does not sound like something that should sit in a driver.

Could we implement this in crypto_engine instead?

Cheers,
--
Email: Herbert Xu <[email protected]>
Home Page: http://gondor.apana.org.au/~herbert/
PGP Key: http://gondor.apana.org.au/~herbert/pubkey.txt

2023-10-20 13:03:17

by Cabiddu, Giovanni

[permalink] [raw]
Subject: Re: [PATCH 00/11] crypto: qat - add rate limiting feature to qat_4xxx

On Fri, Oct 20, 2023 at 01:27:00PM +0800, Herbert Xu wrote:
> On Wed, Oct 11, 2023 at 02:14:58PM +0200, Damian Muszynski wrote:
> > This set introduces the rate limiting feature to the Intel QAT accelerator.
> > The Rate Limiting allows control of the rate of the requests that can be
> > submitted on a ring pair (RP). This allows sharing a QAT device among
> > multiple users while ensuring a guaranteed throughput.
>
> This does not sound like something that should sit in a driver.
To clarify, this is enabling *hardware* rate limiting.

The driver is not doing any active control flow on the jobs submitted to
the accelerator. Instead, it provides a mechanism that allows system
admins to set policies that are then programmed to the device.
Such policies allow to restrict the flow or provide a committed rate on
selected ring pairs.

In brief, the role of the driver is parameter checking and device
configuration based on inputs.

> Could we implement this in crypto_engine instead?
I do not see how this can be implemented in the crypto_engine as rate
limiting functionalities in QAT are applicable not only to requests
submitted through the crypto API but also to requests submitted from a
user space application through qatlib and from a guest virtual machine.

Regards,

--
Giovanni