• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (C) STMicroelectronics SA 2015
4  * Authors: Yannick Fertre <yannick.fertre@st.com>
5  *          Hugues Fruchet <hugues.fruchet@st.com>
6  */
7 
8 #include <linux/clk.h>
9 #include <linux/interrupt.h>
10 #include <linux/platform_device.h>
11 #include <linux/pm_runtime.h>
12 #ifdef CONFIG_VIDEO_STI_HVA_DEBUGFS
13 #include <linux/seq_file.h>
14 #endif
15 
16 #include "hva.h"
17 #include "hva-hw.h"
18 
19 /* HVA register offsets */
20 #define HVA_HIF_REG_RST                 0x0100U
21 #define HVA_HIF_REG_RST_ACK             0x0104U
22 #define HVA_HIF_REG_MIF_CFG             0x0108U
23 #define HVA_HIF_REG_HEC_MIF_CFG         0x010CU
24 #define HVA_HIF_REG_CFL                 0x0110U
25 #define HVA_HIF_FIFO_CMD                0x0114U
26 #define HVA_HIF_FIFO_STS                0x0118U
27 #define HVA_HIF_REG_SFL                 0x011CU
28 #define HVA_HIF_REG_IT_ACK              0x0120U
29 #define HVA_HIF_REG_ERR_IT_ACK          0x0124U
30 #define HVA_HIF_REG_LMI_ERR             0x0128U
31 #define HVA_HIF_REG_EMI_ERR             0x012CU
32 #define HVA_HIF_REG_HEC_MIF_ERR         0x0130U
33 #define HVA_HIF_REG_HEC_STS             0x0134U
34 #define HVA_HIF_REG_HVC_STS             0x0138U
35 #define HVA_HIF_REG_HJE_STS             0x013CU
36 #define HVA_HIF_REG_CNT                 0x0140U
37 #define HVA_HIF_REG_HEC_CHKSYN_DIS      0x0144U
38 #define HVA_HIF_REG_CLK_GATING          0x0148U
39 #define HVA_HIF_REG_VERSION             0x014CU
40 #define HVA_HIF_REG_BSM                 0x0150U
41 
42 /* define value for version id register (HVA_HIF_REG_VERSION) */
43 #define VERSION_ID_MASK	0x0000FFFF
44 
45 /* define values for BSM register (HVA_HIF_REG_BSM) */
46 #define BSM_CFG_VAL1	0x0003F000
47 #define BSM_CFG_VAL2	0x003F0000
48 
49 /* define values for memory interface register (HVA_HIF_REG_MIF_CFG) */
50 #define MIF_CFG_VAL1	0x04460446
51 #define MIF_CFG_VAL2	0x04460806
52 #define MIF_CFG_VAL3	0x00000000
53 
54 /* define value for HEC memory interface register (HVA_HIF_REG_MIF_CFG) */
55 #define HEC_MIF_CFG_VAL	0x000000C4
56 
57 /*  Bits definition for clock gating register (HVA_HIF_REG_CLK_GATING) */
58 #define CLK_GATING_HVC	BIT(0)
59 #define CLK_GATING_HEC	BIT(1)
60 #define CLK_GATING_HJE	BIT(2)
61 
62 /* fix hva clock rate */
63 #define CLK_RATE		300000000
64 
65 /* fix delay for pmruntime */
66 #define AUTOSUSPEND_DELAY_MS	3
67 
68 /*
69  * hw encode error values
70  * NO_ERROR: Success, Task OK
71  * H264_BITSTREAM_OVERSIZE: VECH264 Bitstream size > bitstream buffer
72  * H264_FRAME_SKIPPED: VECH264 Frame skipped (refers to CPB Buffer Size)
73  * H264_SLICE_LIMIT_SIZE: VECH264 MB > slice limit size
74  * H264_MAX_SLICE_NUMBER: VECH264 max slice number reached
75  * H264_SLICE_READY: VECH264 Slice ready
76  * TASK_LIST_FULL: HVA/FPC task list full
77 		   (discard latest transform command)
78  * UNKNOWN_COMMAND: Transform command not known by HVA/FPC
79  * WRONG_CODEC_OR_RESOLUTION: Wrong Codec or Resolution Selection
80  * NO_INT_COMPLETION: Time-out on interrupt completion
81  * LMI_ERR: Local Memory Interface Error
82  * EMI_ERR: External Memory Interface Error
83  * HECMI_ERR: HEC Memory Interface Error
84  */
85 enum hva_hw_error {
86 	NO_ERROR = 0x0,
87 	H264_BITSTREAM_OVERSIZE = 0x2,
88 	H264_FRAME_SKIPPED = 0x4,
89 	H264_SLICE_LIMIT_SIZE = 0x5,
90 	H264_MAX_SLICE_NUMBER = 0x7,
91 	H264_SLICE_READY = 0x8,
92 	TASK_LIST_FULL = 0xF0,
93 	UNKNOWN_COMMAND = 0xF1,
94 	WRONG_CODEC_OR_RESOLUTION = 0xF4,
95 	NO_INT_COMPLETION = 0x100,
96 	LMI_ERR = 0x101,
97 	EMI_ERR = 0x102,
98 	HECMI_ERR = 0x103,
99 };
100 
hva_hw_its_interrupt(int irq,void * data)101 static irqreturn_t hva_hw_its_interrupt(int irq, void *data)
102 {
103 	struct hva_dev *hva = data;
104 
105 	/* read status registers */
106 	hva->sts_reg = readl_relaxed(hva->regs + HVA_HIF_FIFO_STS);
107 	hva->sfl_reg = readl_relaxed(hva->regs + HVA_HIF_REG_SFL);
108 
109 	/* acknowledge interruption */
110 	writel_relaxed(0x1, hva->regs + HVA_HIF_REG_IT_ACK);
111 
112 	return IRQ_WAKE_THREAD;
113 }
114 
hva_hw_its_irq_thread(int irq,void * arg)115 static irqreturn_t hva_hw_its_irq_thread(int irq, void *arg)
116 {
117 	struct hva_dev *hva = arg;
118 	struct device *dev = hva_to_dev(hva);
119 	u32 status = hva->sts_reg & 0xFF;
120 	u8 ctx_id = 0;
121 	struct hva_ctx *ctx = NULL;
122 
123 	dev_dbg(dev, "%s     %s: status: 0x%02x fifo level: 0x%02x\n",
124 		HVA_PREFIX, __func__, hva->sts_reg & 0xFF, hva->sfl_reg & 0xF);
125 
126 	/*
127 	 * status: task_id[31:16] client_id[15:8] status[7:0]
128 	 * the context identifier is retrieved from the client identifier
129 	 */
130 	ctx_id = (hva->sts_reg & 0xFF00) >> 8;
131 	if (ctx_id >= HVA_MAX_INSTANCES) {
132 		dev_err(dev, "%s     %s: bad context identifier: %d\n",
133 			ctx->name, __func__, ctx_id);
134 		ctx->hw_err = true;
135 		goto out;
136 	}
137 
138 	ctx = hva->instances[ctx_id];
139 	if (!ctx)
140 		goto out;
141 
142 	switch (status) {
143 	case NO_ERROR:
144 		dev_dbg(dev, "%s     %s: no error\n",
145 			ctx->name, __func__);
146 		ctx->hw_err = false;
147 		break;
148 	case H264_SLICE_READY:
149 		dev_dbg(dev, "%s     %s: h264 slice ready\n",
150 			ctx->name, __func__);
151 		ctx->hw_err = false;
152 		break;
153 	case H264_FRAME_SKIPPED:
154 		dev_dbg(dev, "%s     %s: h264 frame skipped\n",
155 			ctx->name, __func__);
156 		ctx->hw_err = false;
157 		break;
158 	case H264_BITSTREAM_OVERSIZE:
159 		dev_err(dev, "%s     %s:h264 bitstream oversize\n",
160 			ctx->name, __func__);
161 		ctx->hw_err = true;
162 		break;
163 	case H264_SLICE_LIMIT_SIZE:
164 		dev_err(dev, "%s     %s: h264 slice limit size is reached\n",
165 			ctx->name, __func__);
166 		ctx->hw_err = true;
167 		break;
168 	case H264_MAX_SLICE_NUMBER:
169 		dev_err(dev, "%s     %s: h264 max slice number is reached\n",
170 			ctx->name, __func__);
171 		ctx->hw_err = true;
172 		break;
173 	case TASK_LIST_FULL:
174 		dev_err(dev, "%s     %s:task list full\n",
175 			ctx->name, __func__);
176 		ctx->hw_err = true;
177 		break;
178 	case UNKNOWN_COMMAND:
179 		dev_err(dev, "%s     %s: command not known\n",
180 			ctx->name, __func__);
181 		ctx->hw_err = true;
182 		break;
183 	case WRONG_CODEC_OR_RESOLUTION:
184 		dev_err(dev, "%s     %s: wrong codec or resolution\n",
185 			ctx->name, __func__);
186 		ctx->hw_err = true;
187 		break;
188 	default:
189 		dev_err(dev, "%s     %s: status not recognized\n",
190 			ctx->name, __func__);
191 		ctx->hw_err = true;
192 		break;
193 	}
194 out:
195 	complete(&hva->interrupt);
196 
197 	return IRQ_HANDLED;
198 }
199 
hva_hw_err_interrupt(int irq,void * data)200 static irqreturn_t hva_hw_err_interrupt(int irq, void *data)
201 {
202 	struct hva_dev *hva = data;
203 
204 	/* read status registers */
205 	hva->sts_reg = readl_relaxed(hva->regs + HVA_HIF_FIFO_STS);
206 	hva->sfl_reg = readl_relaxed(hva->regs + HVA_HIF_REG_SFL);
207 
208 	/* read error registers */
209 	hva->lmi_err_reg = readl_relaxed(hva->regs + HVA_HIF_REG_LMI_ERR);
210 	hva->emi_err_reg = readl_relaxed(hva->regs + HVA_HIF_REG_EMI_ERR);
211 	hva->hec_mif_err_reg = readl_relaxed(hva->regs +
212 					     HVA_HIF_REG_HEC_MIF_ERR);
213 
214 	/* acknowledge interruption */
215 	writel_relaxed(0x1, hva->regs + HVA_HIF_REG_IT_ACK);
216 
217 	return IRQ_WAKE_THREAD;
218 }
219 
hva_hw_err_irq_thread(int irq,void * arg)220 static irqreturn_t hva_hw_err_irq_thread(int irq, void *arg)
221 {
222 	struct hva_dev *hva = arg;
223 	struct device *dev = hva_to_dev(hva);
224 	u8 ctx_id = 0;
225 	struct hva_ctx *ctx;
226 
227 	dev_dbg(dev, "%s     status: 0x%02x fifo level: 0x%02x\n",
228 		HVA_PREFIX, hva->sts_reg & 0xFF, hva->sfl_reg & 0xF);
229 
230 	/*
231 	 * status: task_id[31:16] client_id[15:8] status[7:0]
232 	 * the context identifier is retrieved from the client identifier
233 	 */
234 	ctx_id = (hva->sts_reg & 0xFF00) >> 8;
235 	if (ctx_id >= HVA_MAX_INSTANCES) {
236 		dev_err(dev, "%s     bad context identifier: %d\n", HVA_PREFIX,
237 			ctx_id);
238 		goto out;
239 	}
240 
241 	ctx = hva->instances[ctx_id];
242 	if (!ctx)
243 		goto out;
244 
245 	if (hva->lmi_err_reg) {
246 		dev_err(dev, "%s     local memory interface error: 0x%08x\n",
247 			ctx->name, hva->lmi_err_reg);
248 		ctx->hw_err = true;
249 	}
250 
251 	if (hva->emi_err_reg) {
252 		dev_err(dev, "%s     external memory interface error: 0x%08x\n",
253 			ctx->name, hva->emi_err_reg);
254 		ctx->hw_err = true;
255 	}
256 
257 	if (hva->hec_mif_err_reg) {
258 		dev_err(dev, "%s     hec memory interface error: 0x%08x\n",
259 			ctx->name, hva->hec_mif_err_reg);
260 		ctx->hw_err = true;
261 	}
262 out:
263 	complete(&hva->interrupt);
264 
265 	return IRQ_HANDLED;
266 }
267 
hva_hw_get_ip_version(struct hva_dev * hva)268 static unsigned long int hva_hw_get_ip_version(struct hva_dev *hva)
269 {
270 	struct device *dev = hva_to_dev(hva);
271 	unsigned long int version;
272 
273 	if (pm_runtime_get_sync(dev) < 0) {
274 		dev_err(dev, "%s     failed to get pm_runtime\n", HVA_PREFIX);
275 		pm_runtime_put_noidle(dev);
276 		mutex_unlock(&hva->protect_mutex);
277 		return -EFAULT;
278 	}
279 
280 	version = readl_relaxed(hva->regs + HVA_HIF_REG_VERSION) &
281 				VERSION_ID_MASK;
282 
283 	pm_runtime_put_autosuspend(dev);
284 
285 	switch (version) {
286 	case HVA_VERSION_V400:
287 		dev_dbg(dev, "%s     IP hardware version 0x%lx\n",
288 			HVA_PREFIX, version);
289 		break;
290 	default:
291 		dev_err(dev, "%s     unknown IP hardware version 0x%lx\n",
292 			HVA_PREFIX, version);
293 		version = HVA_VERSION_UNKNOWN;
294 		break;
295 	}
296 
297 	return version;
298 }
299 
hva_hw_probe(struct platform_device * pdev,struct hva_dev * hva)300 int hva_hw_probe(struct platform_device *pdev, struct hva_dev *hva)
301 {
302 	struct device *dev = &pdev->dev;
303 	struct resource *regs;
304 	struct resource *esram;
305 	int ret;
306 
307 	WARN_ON(!hva);
308 
309 	/* get memory for registers */
310 	regs = platform_get_resource(pdev, IORESOURCE_MEM, 0);
311 	hva->regs = devm_ioremap_resource(dev, regs);
312 	if (IS_ERR(hva->regs)) {
313 		dev_err(dev, "%s     failed to get regs\n", HVA_PREFIX);
314 		return PTR_ERR(hva->regs);
315 	}
316 
317 	/* get memory for esram */
318 	esram = platform_get_resource(pdev, IORESOURCE_MEM, 1);
319 	if (!esram) {
320 		dev_err(dev, "%s     failed to get esram\n", HVA_PREFIX);
321 		return -ENODEV;
322 	}
323 	hva->esram_addr = esram->start;
324 	hva->esram_size = resource_size(esram);
325 
326 	dev_info(dev, "%s     esram reserved for address: 0x%x size:%d\n",
327 		 HVA_PREFIX, hva->esram_addr, hva->esram_size);
328 
329 	/* get clock resource */
330 	hva->clk = devm_clk_get(dev, "clk_hva");
331 	if (IS_ERR(hva->clk)) {
332 		dev_err(dev, "%s     failed to get clock\n", HVA_PREFIX);
333 		return PTR_ERR(hva->clk);
334 	}
335 
336 	ret = clk_prepare(hva->clk);
337 	if (ret < 0) {
338 		dev_err(dev, "%s     failed to prepare clock\n", HVA_PREFIX);
339 		hva->clk = ERR_PTR(-EINVAL);
340 		return ret;
341 	}
342 
343 	/* get status interruption resource */
344 	ret  = platform_get_irq(pdev, 0);
345 	if (ret < 0) {
346 		dev_err(dev, "%s     failed to get status IRQ\n", HVA_PREFIX);
347 		goto err_clk;
348 	}
349 	hva->irq_its = ret;
350 
351 	ret = devm_request_threaded_irq(dev, hva->irq_its, hva_hw_its_interrupt,
352 					hva_hw_its_irq_thread,
353 					IRQF_ONESHOT,
354 					"hva_its_irq", hva);
355 	if (ret) {
356 		dev_err(dev, "%s     failed to install status IRQ 0x%x\n",
357 			HVA_PREFIX, hva->irq_its);
358 		goto err_clk;
359 	}
360 	disable_irq(hva->irq_its);
361 
362 	/* get error interruption resource */
363 	ret = platform_get_irq(pdev, 1);
364 	if (ret < 0) {
365 		dev_err(dev, "%s     failed to get error IRQ\n", HVA_PREFIX);
366 		goto err_clk;
367 	}
368 	hva->irq_err = ret;
369 
370 	ret = devm_request_threaded_irq(dev, hva->irq_err, hva_hw_err_interrupt,
371 					hva_hw_err_irq_thread,
372 					IRQF_ONESHOT,
373 					"hva_err_irq", hva);
374 	if (ret) {
375 		dev_err(dev, "%s     failed to install error IRQ 0x%x\n",
376 			HVA_PREFIX, hva->irq_err);
377 		goto err_clk;
378 	}
379 	disable_irq(hva->irq_err);
380 
381 	/* initialise protection mutex */
382 	mutex_init(&hva->protect_mutex);
383 
384 	/* initialise completion signal */
385 	init_completion(&hva->interrupt);
386 
387 	/* initialise runtime power management */
388 	pm_runtime_set_autosuspend_delay(dev, AUTOSUSPEND_DELAY_MS);
389 	pm_runtime_use_autosuspend(dev);
390 	pm_runtime_set_suspended(dev);
391 	pm_runtime_enable(dev);
392 
393 	ret = pm_runtime_get_sync(dev);
394 	if (ret < 0) {
395 		dev_err(dev, "%s     failed to set PM\n", HVA_PREFIX);
396 		goto err_pm;
397 	}
398 
399 	/* check IP hardware version */
400 	hva->ip_version = hva_hw_get_ip_version(hva);
401 
402 	if (hva->ip_version == HVA_VERSION_UNKNOWN) {
403 		ret = -EINVAL;
404 		goto err_pm;
405 	}
406 
407 	dev_info(dev, "%s     found hva device (version 0x%lx)\n", HVA_PREFIX,
408 		 hva->ip_version);
409 
410 	return 0;
411 
412 err_pm:
413 	pm_runtime_put(dev);
414 err_clk:
415 	if (hva->clk)
416 		clk_unprepare(hva->clk);
417 
418 	return ret;
419 }
420 
hva_hw_remove(struct hva_dev * hva)421 void hva_hw_remove(struct hva_dev *hva)
422 {
423 	struct device *dev = hva_to_dev(hva);
424 
425 	disable_irq(hva->irq_its);
426 	disable_irq(hva->irq_err);
427 
428 	pm_runtime_put_autosuspend(dev);
429 	pm_runtime_disable(dev);
430 }
431 
hva_hw_runtime_suspend(struct device * dev)432 int hva_hw_runtime_suspend(struct device *dev)
433 {
434 	struct hva_dev *hva = dev_get_drvdata(dev);
435 
436 	clk_disable_unprepare(hva->clk);
437 
438 	return 0;
439 }
440 
hva_hw_runtime_resume(struct device * dev)441 int hva_hw_runtime_resume(struct device *dev)
442 {
443 	struct hva_dev *hva = dev_get_drvdata(dev);
444 
445 	if (clk_prepare_enable(hva->clk)) {
446 		dev_err(hva->dev, "%s     failed to prepare hva clk\n",
447 			HVA_PREFIX);
448 		return -EINVAL;
449 	}
450 
451 	if (clk_set_rate(hva->clk, CLK_RATE)) {
452 		dev_err(dev, "%s     failed to set clock frequency\n",
453 			HVA_PREFIX);
454 		return -EINVAL;
455 	}
456 
457 	return 0;
458 }
459 
hva_hw_execute_task(struct hva_ctx * ctx,enum hva_hw_cmd_type cmd,struct hva_buffer * task)460 int hva_hw_execute_task(struct hva_ctx *ctx, enum hva_hw_cmd_type cmd,
461 			struct hva_buffer *task)
462 {
463 	struct hva_dev *hva = ctx_to_hdev(ctx);
464 	struct device *dev = hva_to_dev(hva);
465 	u8 client_id = ctx->id;
466 	int ret;
467 	u32 reg = 0;
468 
469 	mutex_lock(&hva->protect_mutex);
470 
471 	/* enable irqs */
472 	enable_irq(hva->irq_its);
473 	enable_irq(hva->irq_err);
474 
475 	if (pm_runtime_get_sync(dev) < 0) {
476 		dev_err(dev, "%s     failed to get pm_runtime\n", ctx->name);
477 		ctx->sys_errors++;
478 		ret = -EFAULT;
479 		goto out;
480 	}
481 
482 	reg = readl_relaxed(hva->regs + HVA_HIF_REG_CLK_GATING);
483 	switch (cmd) {
484 	case H264_ENC:
485 		reg |= CLK_GATING_HVC;
486 		break;
487 	default:
488 		dev_dbg(dev, "%s     unknown command 0x%x\n", ctx->name, cmd);
489 		ctx->encode_errors++;
490 		ret = -EFAULT;
491 		goto out;
492 	}
493 	writel_relaxed(reg, hva->regs + HVA_HIF_REG_CLK_GATING);
494 
495 	dev_dbg(dev, "%s     %s: write configuration registers\n", ctx->name,
496 		__func__);
497 
498 	/* byte swap config */
499 	writel_relaxed(BSM_CFG_VAL1, hva->regs + HVA_HIF_REG_BSM);
500 
501 	/* define Max Opcode Size and Max Message Size for LMI and EMI */
502 	writel_relaxed(MIF_CFG_VAL3, hva->regs + HVA_HIF_REG_MIF_CFG);
503 	writel_relaxed(HEC_MIF_CFG_VAL, hva->regs + HVA_HIF_REG_HEC_MIF_CFG);
504 
505 	/*
506 	 * command FIFO: task_id[31:16] client_id[15:8] command_type[7:0]
507 	 * the context identifier is provided as client identifier to the
508 	 * hardware, and is retrieved in the interrupt functions from the
509 	 * status register
510 	 */
511 	dev_dbg(dev, "%s     %s: send task (cmd: %d, task_desc: %pad)\n",
512 		ctx->name, __func__, cmd + (client_id << 8), &task->paddr);
513 	writel_relaxed(cmd + (client_id << 8), hva->regs + HVA_HIF_FIFO_CMD);
514 	writel_relaxed(task->paddr, hva->regs + HVA_HIF_FIFO_CMD);
515 
516 	if (!wait_for_completion_timeout(&hva->interrupt,
517 					 msecs_to_jiffies(2000))) {
518 		dev_err(dev, "%s     %s: time out on completion\n", ctx->name,
519 			__func__);
520 		ctx->encode_errors++;
521 		ret = -EFAULT;
522 		goto out;
523 	}
524 
525 	/* get encoding status */
526 	ret = ctx->hw_err ? -EFAULT : 0;
527 
528 	ctx->encode_errors += ctx->hw_err ? 1 : 0;
529 
530 out:
531 	disable_irq(hva->irq_its);
532 	disable_irq(hva->irq_err);
533 
534 	switch (cmd) {
535 	case H264_ENC:
536 		reg &= ~CLK_GATING_HVC;
537 		writel_relaxed(reg, hva->regs + HVA_HIF_REG_CLK_GATING);
538 		break;
539 	default:
540 		dev_dbg(dev, "%s     unknown command 0x%x\n", ctx->name, cmd);
541 	}
542 
543 	pm_runtime_put_autosuspend(dev);
544 	mutex_unlock(&hva->protect_mutex);
545 
546 	return ret;
547 }
548 
549 #ifdef CONFIG_VIDEO_STI_HVA_DEBUGFS
550 #define DUMP(reg) seq_printf(s, "%-30s: 0x%08X\n",\
551 			     #reg, readl_relaxed(hva->regs + reg))
552 
hva_hw_dump_regs(struct hva_dev * hva,struct seq_file * s)553 void hva_hw_dump_regs(struct hva_dev *hva, struct seq_file *s)
554 {
555 	struct device *dev = hva_to_dev(hva);
556 
557 	mutex_lock(&hva->protect_mutex);
558 
559 	if (pm_runtime_get_sync(dev) < 0) {
560 		seq_puts(s, "Cannot wake up IP\n");
561 		pm_runtime_put_noidle(dev);
562 		mutex_unlock(&hva->protect_mutex);
563 		return;
564 	}
565 
566 	seq_printf(s, "Registers:\nReg @ = 0x%p\n", hva->regs);
567 
568 	DUMP(HVA_HIF_REG_RST);
569 	DUMP(HVA_HIF_REG_RST_ACK);
570 	DUMP(HVA_HIF_REG_MIF_CFG);
571 	DUMP(HVA_HIF_REG_HEC_MIF_CFG);
572 	DUMP(HVA_HIF_REG_CFL);
573 	DUMP(HVA_HIF_REG_SFL);
574 	DUMP(HVA_HIF_REG_LMI_ERR);
575 	DUMP(HVA_HIF_REG_EMI_ERR);
576 	DUMP(HVA_HIF_REG_HEC_MIF_ERR);
577 	DUMP(HVA_HIF_REG_HEC_STS);
578 	DUMP(HVA_HIF_REG_HVC_STS);
579 	DUMP(HVA_HIF_REG_HJE_STS);
580 	DUMP(HVA_HIF_REG_CNT);
581 	DUMP(HVA_HIF_REG_HEC_CHKSYN_DIS);
582 	DUMP(HVA_HIF_REG_CLK_GATING);
583 	DUMP(HVA_HIF_REG_VERSION);
584 
585 	pm_runtime_put_autosuspend(dev);
586 	mutex_unlock(&hva->protect_mutex);
587 }
588 #endif
589