• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // SPDX-License-Identifier: (BSD-3-Clause OR GPL-2.0-only)
2 /* Copyright(c) 2014 - 2020 Intel Corporation */
3 #include <adf_accel_devices.h>
4 #include <adf_pf2vf_msg.h>
5 #include <adf_common_drv.h>
6 #include <adf_gen2_hw_data.h>
7 #include "adf_dh895xcc_hw_data.h"
8 #include "icp_qat_hw.h"
9 
10 /* Worker thread to service arbiter mappings */
11 static const u32 thrd_to_arb_map[ADF_DH895XCC_MAX_ACCELENGINES] = {
12 	0x12222AAA, 0x11666666, 0x12222AAA, 0x11666666,
13 	0x12222AAA, 0x11222222, 0x12222AAA, 0x11222222,
14 	0x12222AAA, 0x11222222, 0x12222AAA, 0x11222222
15 };
16 
17 static struct adf_hw_device_class dh895xcc_class = {
18 	.name = ADF_DH895XCC_DEVICE_NAME,
19 	.type = DEV_DH895XCC,
20 	.instances = 0
21 };
22 
get_accel_mask(struct adf_hw_device_data * self)23 static u32 get_accel_mask(struct adf_hw_device_data *self)
24 {
25 	u32 fuses = self->fuses;
26 
27 	return ~fuses >> ADF_DH895XCC_ACCELERATORS_REG_OFFSET &
28 			 ADF_DH895XCC_ACCELERATORS_MASK;
29 }
30 
get_ae_mask(struct adf_hw_device_data * self)31 static u32 get_ae_mask(struct adf_hw_device_data *self)
32 {
33 	u32 fuses = self->fuses;
34 
35 	return ~fuses & ADF_DH895XCC_ACCELENGINES_MASK;
36 }
37 
get_num_accels(struct adf_hw_device_data * self)38 static u32 get_num_accels(struct adf_hw_device_data *self)
39 {
40 	u32 i, ctr = 0;
41 
42 	if (!self || !self->accel_mask)
43 		return 0;
44 
45 	for (i = 0; i < ADF_DH895XCC_MAX_ACCELERATORS; i++) {
46 		if (self->accel_mask & (1 << i))
47 			ctr++;
48 	}
49 	return ctr;
50 }
51 
get_num_aes(struct adf_hw_device_data * self)52 static u32 get_num_aes(struct adf_hw_device_data *self)
53 {
54 	u32 i, ctr = 0;
55 
56 	if (!self || !self->ae_mask)
57 		return 0;
58 
59 	for (i = 0; i < ADF_DH895XCC_MAX_ACCELENGINES; i++) {
60 		if (self->ae_mask & (1 << i))
61 			ctr++;
62 	}
63 	return ctr;
64 }
65 
get_misc_bar_id(struct adf_hw_device_data * self)66 static u32 get_misc_bar_id(struct adf_hw_device_data *self)
67 {
68 	return ADF_DH895XCC_PMISC_BAR;
69 }
70 
get_etr_bar_id(struct adf_hw_device_data * self)71 static u32 get_etr_bar_id(struct adf_hw_device_data *self)
72 {
73 	return ADF_DH895XCC_ETR_BAR;
74 }
75 
get_sram_bar_id(struct adf_hw_device_data * self)76 static u32 get_sram_bar_id(struct adf_hw_device_data *self)
77 {
78 	return ADF_DH895XCC_SRAM_BAR;
79 }
80 
get_accel_cap(struct adf_accel_dev * accel_dev)81 static u32 get_accel_cap(struct adf_accel_dev *accel_dev)
82 {
83 	struct pci_dev *pdev = accel_dev->accel_pci_dev.pci_dev;
84 	u32 capabilities;
85 	u32 legfuses;
86 
87 	capabilities = ICP_ACCEL_CAPABILITIES_CRYPTO_SYMMETRIC |
88 		       ICP_ACCEL_CAPABILITIES_CRYPTO_ASYMMETRIC |
89 		       ICP_ACCEL_CAPABILITIES_AUTHENTICATION |
90 		       ICP_ACCEL_CAPABILITIES_CIPHER |
91 		       ICP_ACCEL_CAPABILITIES_COMPRESSION;
92 
93 	/* Read accelerator capabilities mask */
94 	pci_read_config_dword(pdev, ADF_DEVICE_LEGFUSE_OFFSET, &legfuses);
95 
96 	/* A set bit in legfuses means the feature is OFF in this SKU */
97 	if (legfuses & ICP_ACCEL_MASK_CIPHER_SLICE) {
98 		capabilities &= ~ICP_ACCEL_CAPABILITIES_CRYPTO_SYMMETRIC;
99 		capabilities &= ~ICP_ACCEL_CAPABILITIES_CIPHER;
100 	}
101 	if (legfuses & ICP_ACCEL_MASK_PKE_SLICE)
102 		capabilities &= ~ICP_ACCEL_CAPABILITIES_CRYPTO_ASYMMETRIC;
103 	if (legfuses & ICP_ACCEL_MASK_AUTH_SLICE) {
104 		capabilities &= ~ICP_ACCEL_CAPABILITIES_AUTHENTICATION;
105 		capabilities &= ~ICP_ACCEL_CAPABILITIES_CIPHER;
106 	}
107 	if (legfuses & ICP_ACCEL_MASK_COMPRESS_SLICE)
108 		capabilities &= ~ICP_ACCEL_CAPABILITIES_COMPRESSION;
109 
110 	return capabilities;
111 }
112 
get_sku(struct adf_hw_device_data * self)113 static enum dev_sku_info get_sku(struct adf_hw_device_data *self)
114 {
115 	int sku = (self->fuses & ADF_DH895XCC_FUSECTL_SKU_MASK)
116 	    >> ADF_DH895XCC_FUSECTL_SKU_SHIFT;
117 
118 	switch (sku) {
119 	case ADF_DH895XCC_FUSECTL_SKU_1:
120 		return DEV_SKU_1;
121 	case ADF_DH895XCC_FUSECTL_SKU_2:
122 		return DEV_SKU_2;
123 	case ADF_DH895XCC_FUSECTL_SKU_3:
124 		return DEV_SKU_3;
125 	case ADF_DH895XCC_FUSECTL_SKU_4:
126 		return DEV_SKU_4;
127 	default:
128 		return DEV_SKU_UNKNOWN;
129 	}
130 	return DEV_SKU_UNKNOWN;
131 }
132 
adf_get_arbiter_mapping(void)133 static const u32 *adf_get_arbiter_mapping(void)
134 {
135 	return thrd_to_arb_map;
136 }
137 
get_pf2vf_offset(u32 i)138 static u32 get_pf2vf_offset(u32 i)
139 {
140 	return ADF_DH895XCC_PF2VF_OFFSET(i);
141 }
142 
adf_enable_error_correction(struct adf_accel_dev * accel_dev)143 static void adf_enable_error_correction(struct adf_accel_dev *accel_dev)
144 {
145 	struct adf_hw_device_data *hw_device = accel_dev->hw_device;
146 	struct adf_bar *misc_bar = &GET_BARS(accel_dev)[ADF_DH895XCC_PMISC_BAR];
147 	unsigned long accel_mask = hw_device->accel_mask;
148 	unsigned long ae_mask = hw_device->ae_mask;
149 	void __iomem *csr = misc_bar->virt_addr;
150 	unsigned int val, i;
151 
152 	/* Enable Accel Engine error detection & correction */
153 	for_each_set_bit(i, &ae_mask, GET_MAX_ACCELENGINES(accel_dev)) {
154 		val = ADF_CSR_RD(csr, ADF_DH895XCC_AE_CTX_ENABLES(i));
155 		val |= ADF_DH895XCC_ENABLE_AE_ECC_ERR;
156 		ADF_CSR_WR(csr, ADF_DH895XCC_AE_CTX_ENABLES(i), val);
157 		val = ADF_CSR_RD(csr, ADF_DH895XCC_AE_MISC_CONTROL(i));
158 		val |= ADF_DH895XCC_ENABLE_AE_ECC_PARITY_CORR;
159 		ADF_CSR_WR(csr, ADF_DH895XCC_AE_MISC_CONTROL(i), val);
160 	}
161 
162 	/* Enable shared memory error detection & correction */
163 	for_each_set_bit(i, &accel_mask, ADF_DH895XCC_MAX_ACCELERATORS) {
164 		val = ADF_CSR_RD(csr, ADF_DH895XCC_UERRSSMSH(i));
165 		val |= ADF_DH895XCC_ERRSSMSH_EN;
166 		ADF_CSR_WR(csr, ADF_DH895XCC_UERRSSMSH(i), val);
167 		val = ADF_CSR_RD(csr, ADF_DH895XCC_CERRSSMSH(i));
168 		val |= ADF_DH895XCC_ERRSSMSH_EN;
169 		ADF_CSR_WR(csr, ADF_DH895XCC_CERRSSMSH(i), val);
170 	}
171 }
172 
adf_enable_ints(struct adf_accel_dev * accel_dev)173 static void adf_enable_ints(struct adf_accel_dev *accel_dev)
174 {
175 	void __iomem *addr;
176 
177 	addr = (&GET_BARS(accel_dev)[ADF_DH895XCC_PMISC_BAR])->virt_addr;
178 
179 	/* Enable bundle and misc interrupts */
180 	ADF_CSR_WR(addr, ADF_DH895XCC_SMIAPF0_MASK_OFFSET,
181 		   accel_dev->pf.vf_info ? 0 :
182 			BIT_ULL(GET_MAX_BANKS(accel_dev)) - 1);
183 	ADF_CSR_WR(addr, ADF_DH895XCC_SMIAPF1_MASK_OFFSET,
184 		   ADF_DH895XCC_SMIA1_MASK);
185 }
186 
adf_enable_pf2vf_comms(struct adf_accel_dev * accel_dev)187 static int adf_enable_pf2vf_comms(struct adf_accel_dev *accel_dev)
188 {
189 	spin_lock_init(&accel_dev->pf.vf2pf_ints_lock);
190 
191 	return 0;
192 }
193 
configure_iov_threads(struct adf_accel_dev * accel_dev,bool enable)194 static void configure_iov_threads(struct adf_accel_dev *accel_dev, bool enable)
195 {
196 	adf_gen2_cfg_iov_thds(accel_dev, enable,
197 			      ADF_DH895XCC_AE2FUNC_MAP_GRP_A_NUM_REGS,
198 			      ADF_DH895XCC_AE2FUNC_MAP_GRP_B_NUM_REGS);
199 }
200 
adf_init_hw_data_dh895xcc(struct adf_hw_device_data * hw_data)201 void adf_init_hw_data_dh895xcc(struct adf_hw_device_data *hw_data)
202 {
203 	hw_data->dev_class = &dh895xcc_class;
204 	hw_data->instance_id = dh895xcc_class.instances++;
205 	hw_data->num_banks = ADF_DH895XCC_ETR_MAX_BANKS;
206 	hw_data->num_rings_per_bank = ADF_ETR_MAX_RINGS_PER_BANK;
207 	hw_data->num_accel = ADF_DH895XCC_MAX_ACCELERATORS;
208 	hw_data->num_logical_accel = 1;
209 	hw_data->num_engines = ADF_DH895XCC_MAX_ACCELENGINES;
210 	hw_data->tx_rx_gap = ADF_DH895XCC_RX_RINGS_OFFSET;
211 	hw_data->tx_rings_mask = ADF_DH895XCC_TX_RINGS_MASK;
212 	hw_data->alloc_irq = adf_isr_resource_alloc;
213 	hw_data->free_irq = adf_isr_resource_free;
214 	hw_data->enable_error_correction = adf_enable_error_correction;
215 	hw_data->get_accel_mask = get_accel_mask;
216 	hw_data->get_ae_mask = get_ae_mask;
217 	hw_data->get_accel_cap = get_accel_cap;
218 	hw_data->get_num_accels = get_num_accels;
219 	hw_data->get_num_aes = get_num_aes;
220 	hw_data->get_etr_bar_id = get_etr_bar_id;
221 	hw_data->get_misc_bar_id = get_misc_bar_id;
222 	hw_data->get_admin_info = adf_gen2_get_admin_info;
223 	hw_data->get_arb_info = adf_gen2_get_arb_info;
224 	hw_data->get_sram_bar_id = get_sram_bar_id;
225 	hw_data->get_sku = get_sku;
226 	hw_data->fw_name = ADF_DH895XCC_FW;
227 	hw_data->fw_mmp_name = ADF_DH895XCC_MMP;
228 	hw_data->init_admin_comms = adf_init_admin_comms;
229 	hw_data->exit_admin_comms = adf_exit_admin_comms;
230 	hw_data->configure_iov_threads = configure_iov_threads;
231 	hw_data->send_admin_init = adf_send_admin_init;
232 	hw_data->init_arb = adf_init_arb;
233 	hw_data->exit_arb = adf_exit_arb;
234 	hw_data->get_arb_mapping = adf_get_arbiter_mapping;
235 	hw_data->enable_ints = adf_enable_ints;
236 	hw_data->reset_device = adf_reset_sbr;
237 	hw_data->get_pf2vf_offset = get_pf2vf_offset;
238 	hw_data->enable_pfvf_comms = adf_enable_pf2vf_comms;
239 	hw_data->disable_iov = adf_disable_sriov;
240 	hw_data->min_iov_compat_ver = ADF_PFVF_COMPAT_THIS_VERSION;
241 
242 	adf_gen2_init_hw_csr_ops(&hw_data->csr_ops);
243 }
244 
adf_clean_hw_data_dh895xcc(struct adf_hw_device_data * hw_data)245 void adf_clean_hw_data_dh895xcc(struct adf_hw_device_data *hw_data)
246 {
247 	hw_data->dev_class->instances--;
248 }
249