1 /*
2 * Adaptec AAC series RAID controller driver
3 * (c) Copyright 2001 Red Hat Inc.
4 *
5 * based on the old aacraid driver that is..
6 * Adaptec aacraid device driver for Linux.
7 *
8 * Copyright (c) 2000-2010 Adaptec, Inc.
9 * 2010-2015 PMC-Sierra, Inc. (aacraid@pmc-sierra.com)
10 * 2016-2017 Microsemi Corp. (aacraid@microsemi.com)
11 *
12 * This program is free software; you can redistribute it and/or modify
13 * it under the terms of the GNU General Public License as published by
14 * the Free Software Foundation; either version 2, or (at your option)
15 * any later version.
16 *
17 * This program is distributed in the hope that it will be useful,
18 * but WITHOUT ANY WARRANTY; without even the implied warranty of
19 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
20 * GNU General Public License for more details.
21 *
22 * You should have received a copy of the GNU General Public License
23 * along with this program; see the file COPYING. If not, write to
24 * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA.
25 *
26 * Module Name:
27 * src.c
28 *
29 * Abstract: Hardware Device Interface for PMC SRC based controllers
30 *
31 */
32
33 #include <linux/kernel.h>
34 #include <linux/init.h>
35 #include <linux/types.h>
36 #include <linux/pci.h>
37 #include <linux/spinlock.h>
38 #include <linux/slab.h>
39 #include <linux/blkdev.h>
40 #include <linux/delay.h>
41 #include <linux/completion.h>
42 #include <linux/time.h>
43 #include <linux/interrupt.h>
44 #include <scsi/scsi_host.h>
45
46 #include "aacraid.h"
47
48 static int aac_src_get_sync_status(struct aac_dev *dev);
49
aac_src_intr_message(int irq,void * dev_id)50 static irqreturn_t aac_src_intr_message(int irq, void *dev_id)
51 {
52 struct aac_msix_ctx *ctx;
53 struct aac_dev *dev;
54 unsigned long bellbits, bellbits_shifted;
55 int vector_no;
56 int isFastResponse, mode;
57 u32 index, handle;
58
59 ctx = (struct aac_msix_ctx *)dev_id;
60 dev = ctx->dev;
61 vector_no = ctx->vector_no;
62
63 if (dev->msi_enabled) {
64 mode = AAC_INT_MODE_MSI;
65 if (vector_no == 0) {
66 bellbits = src_readl(dev, MUnit.ODR_MSI);
67 if (bellbits & 0x40000)
68 mode |= AAC_INT_MODE_AIF;
69 if (bellbits & 0x1000)
70 mode |= AAC_INT_MODE_SYNC;
71 }
72 } else {
73 mode = AAC_INT_MODE_INTX;
74 bellbits = src_readl(dev, MUnit.ODR_R);
75 if (bellbits & PmDoorBellResponseSent) {
76 bellbits = PmDoorBellResponseSent;
77 src_writel(dev, MUnit.ODR_C, bellbits);
78 src_readl(dev, MUnit.ODR_C);
79 } else {
80 bellbits_shifted = (bellbits >> SRC_ODR_SHIFT);
81 src_writel(dev, MUnit.ODR_C, bellbits);
82 src_readl(dev, MUnit.ODR_C);
83
84 if (bellbits_shifted & DoorBellAifPending)
85 mode |= AAC_INT_MODE_AIF;
86 else if (bellbits_shifted & OUTBOUNDDOORBELL_0)
87 mode |= AAC_INT_MODE_SYNC;
88 }
89 }
90
91 if (mode & AAC_INT_MODE_SYNC) {
92 unsigned long sflags;
93 struct list_head *entry;
94 int send_it = 0;
95 extern int aac_sync_mode;
96
97 if (!aac_sync_mode && !dev->msi_enabled) {
98 src_writel(dev, MUnit.ODR_C, bellbits);
99 src_readl(dev, MUnit.ODR_C);
100 }
101
102 if (dev->sync_fib) {
103 if (dev->sync_fib->callback)
104 dev->sync_fib->callback(dev->sync_fib->callback_data,
105 dev->sync_fib);
106 spin_lock_irqsave(&dev->sync_fib->event_lock, sflags);
107 if (dev->sync_fib->flags & FIB_CONTEXT_FLAG_WAIT) {
108 dev->management_fib_count--;
109 up(&dev->sync_fib->event_wait);
110 }
111 spin_unlock_irqrestore(&dev->sync_fib->event_lock,
112 sflags);
113 spin_lock_irqsave(&dev->sync_lock, sflags);
114 if (!list_empty(&dev->sync_fib_list)) {
115 entry = dev->sync_fib_list.next;
116 dev->sync_fib = list_entry(entry,
117 struct fib,
118 fiblink);
119 list_del(entry);
120 send_it = 1;
121 } else {
122 dev->sync_fib = NULL;
123 }
124 spin_unlock_irqrestore(&dev->sync_lock, sflags);
125 if (send_it) {
126 aac_adapter_sync_cmd(dev, SEND_SYNCHRONOUS_FIB,
127 (u32)dev->sync_fib->hw_fib_pa,
128 0, 0, 0, 0, 0,
129 NULL, NULL, NULL, NULL, NULL);
130 }
131 }
132 if (!dev->msi_enabled)
133 mode = 0;
134
135 }
136
137 if (mode & AAC_INT_MODE_AIF) {
138 /* handle AIF */
139 if (dev->sa_firmware) {
140 u32 events = src_readl(dev, MUnit.SCR0);
141
142 aac_intr_normal(dev, events, 1, 0, NULL);
143 writel(events, &dev->IndexRegs->Mailbox[0]);
144 src_writel(dev, MUnit.IDR, 1 << 23);
145 } else {
146 if (dev->aif_thread && dev->fsa_dev)
147 aac_intr_normal(dev, 0, 2, 0, NULL);
148 }
149 if (dev->msi_enabled)
150 aac_src_access_devreg(dev, AAC_CLEAR_AIF_BIT);
151 mode = 0;
152 }
153
154 if (mode) {
155 index = dev->host_rrq_idx[vector_no];
156
157 for (;;) {
158 isFastResponse = 0;
159 /* remove toggle bit (31) */
160 handle = le32_to_cpu((dev->host_rrq[index])
161 & 0x7fffffff);
162 /* check fast response bits (30, 1) */
163 if (handle & 0x40000000)
164 isFastResponse = 1;
165 handle &= 0x0000ffff;
166 if (handle == 0)
167 break;
168 handle >>= 2;
169 if (dev->msi_enabled && dev->max_msix > 1)
170 atomic_dec(&dev->rrq_outstanding[vector_no]);
171 aac_intr_normal(dev, handle, 0, isFastResponse, NULL);
172 dev->host_rrq[index++] = 0;
173 if (index == (vector_no + 1) * dev->vector_cap)
174 index = vector_no * dev->vector_cap;
175 dev->host_rrq_idx[vector_no] = index;
176 }
177 mode = 0;
178 }
179
180 return IRQ_HANDLED;
181 }
182
183 /**
184 * aac_src_disable_interrupt - Disable interrupts
185 * @dev: Adapter
186 */
187
aac_src_disable_interrupt(struct aac_dev * dev)188 static void aac_src_disable_interrupt(struct aac_dev *dev)
189 {
190 src_writel(dev, MUnit.OIMR, dev->OIMR = 0xffffffff);
191 }
192
193 /**
194 * aac_src_enable_interrupt_message - Enable interrupts
195 * @dev: Adapter
196 */
197
aac_src_enable_interrupt_message(struct aac_dev * dev)198 static void aac_src_enable_interrupt_message(struct aac_dev *dev)
199 {
200 aac_src_access_devreg(dev, AAC_ENABLE_INTERRUPT);
201 }
202
203 /**
204 * src_sync_cmd - send a command and wait
205 * @dev: Adapter
206 * @command: Command to execute
207 * @p1: first parameter
208 * @ret: adapter status
209 *
210 * This routine will send a synchronous command to the adapter and wait
211 * for its completion.
212 */
213
src_sync_cmd(struct aac_dev * dev,u32 command,u32 p1,u32 p2,u32 p3,u32 p4,u32 p5,u32 p6,u32 * status,u32 * r1,u32 * r2,u32 * r3,u32 * r4)214 static int src_sync_cmd(struct aac_dev *dev, u32 command,
215 u32 p1, u32 p2, u32 p3, u32 p4, u32 p5, u32 p6,
216 u32 *status, u32 * r1, u32 * r2, u32 * r3, u32 * r4)
217 {
218 unsigned long start;
219 unsigned long delay;
220 int ok;
221
222 /*
223 * Write the command into Mailbox 0
224 */
225 writel(command, &dev->IndexRegs->Mailbox[0]);
226 /*
227 * Write the parameters into Mailboxes 1 - 6
228 */
229 writel(p1, &dev->IndexRegs->Mailbox[1]);
230 writel(p2, &dev->IndexRegs->Mailbox[2]);
231 writel(p3, &dev->IndexRegs->Mailbox[3]);
232 writel(p4, &dev->IndexRegs->Mailbox[4]);
233
234 /*
235 * Clear the synch command doorbell to start on a clean slate.
236 */
237 if (!dev->msi_enabled)
238 src_writel(dev,
239 MUnit.ODR_C,
240 OUTBOUNDDOORBELL_0 << SRC_ODR_SHIFT);
241
242 /*
243 * Disable doorbell interrupts
244 */
245 src_writel(dev, MUnit.OIMR, dev->OIMR = 0xffffffff);
246
247 /*
248 * Force the completion of the mask register write before issuing
249 * the interrupt.
250 */
251 src_readl(dev, MUnit.OIMR);
252
253 /*
254 * Signal that there is a new synch command
255 */
256 src_writel(dev, MUnit.IDR, INBOUNDDOORBELL_0 << SRC_IDR_SHIFT);
257
258 if (!dev->sync_mode || command != SEND_SYNCHRONOUS_FIB) {
259 ok = 0;
260 start = jiffies;
261
262 if (command == IOP_RESET_ALWAYS) {
263 /* Wait up to 10 sec */
264 delay = 10*HZ;
265 } else {
266 /* Wait up to 5 minutes */
267 delay = 300*HZ;
268 }
269 while (time_before(jiffies, start+delay)) {
270 udelay(5); /* Delay 5 microseconds to let Mon960 get info. */
271 /*
272 * Mon960 will set doorbell0 bit when it has completed the command.
273 */
274 if (aac_src_get_sync_status(dev) & OUTBOUNDDOORBELL_0) {
275 /*
276 * Clear the doorbell.
277 */
278 if (dev->msi_enabled)
279 aac_src_access_devreg(dev,
280 AAC_CLEAR_SYNC_BIT);
281 else
282 src_writel(dev,
283 MUnit.ODR_C,
284 OUTBOUNDDOORBELL_0 << SRC_ODR_SHIFT);
285 ok = 1;
286 break;
287 }
288 /*
289 * Yield the processor in case we are slow
290 */
291 msleep(1);
292 }
293 if (unlikely(ok != 1)) {
294 /*
295 * Restore interrupt mask even though we timed out
296 */
297 aac_adapter_enable_int(dev);
298 return -ETIMEDOUT;
299 }
300 /*
301 * Pull the synch status from Mailbox 0.
302 */
303 if (status)
304 *status = readl(&dev->IndexRegs->Mailbox[0]);
305 if (r1)
306 *r1 = readl(&dev->IndexRegs->Mailbox[1]);
307 if (r2)
308 *r2 = readl(&dev->IndexRegs->Mailbox[2]);
309 if (r3)
310 *r3 = readl(&dev->IndexRegs->Mailbox[3]);
311 if (r4)
312 *r4 = readl(&dev->IndexRegs->Mailbox[4]);
313 if (command == GET_COMM_PREFERRED_SETTINGS)
314 dev->max_msix =
315 readl(&dev->IndexRegs->Mailbox[5]) & 0xFFFF;
316 /*
317 * Clear the synch command doorbell.
318 */
319 if (!dev->msi_enabled)
320 src_writel(dev,
321 MUnit.ODR_C,
322 OUTBOUNDDOORBELL_0 << SRC_ODR_SHIFT);
323 }
324
325 /*
326 * Restore interrupt mask
327 */
328 aac_adapter_enable_int(dev);
329 return 0;
330 }
331
332 /**
333 * aac_src_interrupt_adapter - interrupt adapter
334 * @dev: Adapter
335 *
336 * Send an interrupt to the i960 and breakpoint it.
337 */
338
aac_src_interrupt_adapter(struct aac_dev * dev)339 static void aac_src_interrupt_adapter(struct aac_dev *dev)
340 {
341 src_sync_cmd(dev, BREAKPOINT_REQUEST,
342 0, 0, 0, 0, 0, 0,
343 NULL, NULL, NULL, NULL, NULL);
344 }
345
346 /**
347 * aac_src_notify_adapter - send an event to the adapter
348 * @dev: Adapter
349 * @event: Event to send
350 *
351 * Notify the i960 that something it probably cares about has
352 * happened.
353 */
354
aac_src_notify_adapter(struct aac_dev * dev,u32 event)355 static void aac_src_notify_adapter(struct aac_dev *dev, u32 event)
356 {
357 switch (event) {
358
359 case AdapNormCmdQue:
360 src_writel(dev, MUnit.ODR_C,
361 INBOUNDDOORBELL_1 << SRC_ODR_SHIFT);
362 break;
363 case HostNormRespNotFull:
364 src_writel(dev, MUnit.ODR_C,
365 INBOUNDDOORBELL_4 << SRC_ODR_SHIFT);
366 break;
367 case AdapNormRespQue:
368 src_writel(dev, MUnit.ODR_C,
369 INBOUNDDOORBELL_2 << SRC_ODR_SHIFT);
370 break;
371 case HostNormCmdNotFull:
372 src_writel(dev, MUnit.ODR_C,
373 INBOUNDDOORBELL_3 << SRC_ODR_SHIFT);
374 break;
375 case FastIo:
376 src_writel(dev, MUnit.ODR_C,
377 INBOUNDDOORBELL_6 << SRC_ODR_SHIFT);
378 break;
379 case AdapPrintfDone:
380 src_writel(dev, MUnit.ODR_C,
381 INBOUNDDOORBELL_5 << SRC_ODR_SHIFT);
382 break;
383 default:
384 BUG();
385 break;
386 }
387 }
388
389 /**
390 * aac_src_start_adapter - activate adapter
391 * @dev: Adapter
392 *
393 * Start up processing on an i960 based AAC adapter
394 */
395
aac_src_start_adapter(struct aac_dev * dev)396 static void aac_src_start_adapter(struct aac_dev *dev)
397 {
398 union aac_init *init;
399 int i;
400
401 /* reset host_rrq_idx first */
402 for (i = 0; i < dev->max_msix; i++) {
403 dev->host_rrq_idx[i] = i * dev->vector_cap;
404 atomic_set(&dev->rrq_outstanding[i], 0);
405 }
406 atomic_set(&dev->msix_counter, 0);
407 dev->fibs_pushed_no = 0;
408
409 init = dev->init;
410 if (dev->comm_interface == AAC_COMM_MESSAGE_TYPE3) {
411 init->r8.host_elapsed_seconds = cpu_to_le32(get_seconds());
412 src_sync_cmd(dev, INIT_STRUCT_BASE_ADDRESS,
413 lower_32_bits(dev->init_pa),
414 upper_32_bits(dev->init_pa),
415 sizeof(struct _r8) +
416 (AAC_MAX_HRRQ - 1) * sizeof(struct _rrq),
417 0, 0, 0, NULL, NULL, NULL, NULL, NULL);
418 } else {
419 init->r7.host_elapsed_seconds = cpu_to_le32(get_seconds());
420 // We can only use a 32 bit address here
421 src_sync_cmd(dev, INIT_STRUCT_BASE_ADDRESS,
422 (u32)(ulong)dev->init_pa, 0, 0, 0, 0, 0,
423 NULL, NULL, NULL, NULL, NULL);
424 }
425
426 }
427
428 /**
429 * aac_src_check_health
430 * @dev: device to check if healthy
431 *
432 * Will attempt to determine if the specified adapter is alive and
433 * capable of handling requests, returning 0 if alive.
434 */
aac_src_check_health(struct aac_dev * dev)435 static int aac_src_check_health(struct aac_dev *dev)
436 {
437 u32 status = src_readl(dev, MUnit.OMR);
438
439 /*
440 * Check to see if the board panic'd.
441 */
442 if (unlikely(status & KERNEL_PANIC))
443 goto err_blink;
444
445 /*
446 * Check to see if the board failed any self tests.
447 */
448 if (unlikely(status & SELF_TEST_FAILED))
449 goto err_out;
450
451 /*
452 * Check to see if the board failed any self tests.
453 */
454 if (unlikely(status & MONITOR_PANIC))
455 goto err_out;
456
457 /*
458 * Wait for the adapter to be up and running.
459 */
460 if (unlikely(!(status & KERNEL_UP_AND_RUNNING)))
461 return -3;
462 /*
463 * Everything is OK
464 */
465 return 0;
466
467 err_out:
468 return -1;
469
470 err_blink:
471 return (status >> 16) & 0xFF;
472 }
473
aac_get_vector(struct aac_dev * dev)474 static inline u32 aac_get_vector(struct aac_dev *dev)
475 {
476 return atomic_inc_return(&dev->msix_counter)%dev->max_msix;
477 }
478
479 /**
480 * aac_src_deliver_message
481 * @fib: fib to issue
482 *
483 * Will send a fib, returning 0 if successful.
484 */
aac_src_deliver_message(struct fib * fib)485 static int aac_src_deliver_message(struct fib *fib)
486 {
487 struct aac_dev *dev = fib->dev;
488 struct aac_queue *q = &dev->queues->queue[AdapNormCmdQueue];
489 u32 fibsize;
490 dma_addr_t address;
491 struct aac_fib_xporthdr *pFibX;
492 int native_hba;
493 #if !defined(writeq)
494 unsigned long flags;
495 #endif
496
497 u16 vector_no;
498
499 atomic_inc(&q->numpending);
500
501 native_hba = (fib->flags & FIB_CONTEXT_FLAG_NATIVE_HBA) ? 1 : 0;
502
503
504 if (dev->msi_enabled && dev->max_msix > 1 &&
505 (native_hba || fib->hw_fib_va->header.Command != AifRequest)) {
506
507 if ((dev->comm_interface == AAC_COMM_MESSAGE_TYPE3)
508 && dev->sa_firmware)
509 vector_no = aac_get_vector(dev);
510 else
511 vector_no = fib->vector_no;
512
513 if (native_hba) {
514 if (fib->flags & FIB_CONTEXT_FLAG_NATIVE_HBA_TMF) {
515 struct aac_hba_tm_req *tm_req;
516
517 tm_req = (struct aac_hba_tm_req *)
518 fib->hw_fib_va;
519 if (tm_req->iu_type ==
520 HBA_IU_TYPE_SCSI_TM_REQ) {
521 ((struct aac_hba_tm_req *)
522 fib->hw_fib_va)->reply_qid
523 = vector_no;
524 ((struct aac_hba_tm_req *)
525 fib->hw_fib_va)->request_id
526 += (vector_no << 16);
527 } else {
528 ((struct aac_hba_reset_req *)
529 fib->hw_fib_va)->reply_qid
530 = vector_no;
531 ((struct aac_hba_reset_req *)
532 fib->hw_fib_va)->request_id
533 += (vector_no << 16);
534 }
535 } else {
536 ((struct aac_hba_cmd_req *)
537 fib->hw_fib_va)->reply_qid
538 = vector_no;
539 ((struct aac_hba_cmd_req *)
540 fib->hw_fib_va)->request_id
541 += (vector_no << 16);
542 }
543 } else {
544 fib->hw_fib_va->header.Handle += (vector_no << 16);
545 }
546 } else {
547 vector_no = 0;
548 }
549
550 atomic_inc(&dev->rrq_outstanding[vector_no]);
551
552 if (native_hba) {
553 address = fib->hw_fib_pa;
554 fibsize = (fib->hbacmd_size + 127) / 128 - 1;
555 if (fibsize > 31)
556 fibsize = 31;
557 address |= fibsize;
558 #if defined(writeq)
559 src_writeq(dev, MUnit.IQN_L, (u64)address);
560 #else
561 spin_lock_irqsave(&fib->dev->iq_lock, flags);
562 src_writel(dev, MUnit.IQN_H,
563 upper_32_bits(address) & 0xffffffff);
564 src_writel(dev, MUnit.IQN_L, address & 0xffffffff);
565 spin_unlock_irqrestore(&fib->dev->iq_lock, flags);
566 #endif
567 } else {
568 if (dev->comm_interface == AAC_COMM_MESSAGE_TYPE2 ||
569 dev->comm_interface == AAC_COMM_MESSAGE_TYPE3) {
570 /* Calculate the amount to the fibsize bits */
571 fibsize = (le16_to_cpu(fib->hw_fib_va->header.Size)
572 + 127) / 128 - 1;
573 /* New FIB header, 32-bit */
574 address = fib->hw_fib_pa;
575 fib->hw_fib_va->header.StructType = FIB_MAGIC2;
576 fib->hw_fib_va->header.SenderFibAddress =
577 cpu_to_le32((u32)address);
578 fib->hw_fib_va->header.u.TimeStamp = 0;
579 WARN_ON(upper_32_bits(address) != 0L);
580 } else {
581 /* Calculate the amount to the fibsize bits */
582 fibsize = (sizeof(struct aac_fib_xporthdr) +
583 le16_to_cpu(fib->hw_fib_va->header.Size)
584 + 127) / 128 - 1;
585 /* Fill XPORT header */
586 pFibX = (struct aac_fib_xporthdr *)
587 ((unsigned char *)fib->hw_fib_va -
588 sizeof(struct aac_fib_xporthdr));
589 pFibX->Handle = fib->hw_fib_va->header.Handle;
590 pFibX->HostAddress =
591 cpu_to_le64((u64)fib->hw_fib_pa);
592 pFibX->Size = cpu_to_le32(
593 le16_to_cpu(fib->hw_fib_va->header.Size));
594 address = fib->hw_fib_pa -
595 (u64)sizeof(struct aac_fib_xporthdr);
596 }
597 if (fibsize > 31)
598 fibsize = 31;
599 address |= fibsize;
600
601 #if defined(writeq)
602 src_writeq(dev, MUnit.IQ_L, (u64)address);
603 #else
604 spin_lock_irqsave(&fib->dev->iq_lock, flags);
605 src_writel(dev, MUnit.IQ_H,
606 upper_32_bits(address) & 0xffffffff);
607 src_writel(dev, MUnit.IQ_L, address & 0xffffffff);
608 spin_unlock_irqrestore(&fib->dev->iq_lock, flags);
609 #endif
610 }
611 return 0;
612 }
613
614 /**
615 * aac_src_ioremap
616 * @size: mapping resize request
617 *
618 */
aac_src_ioremap(struct aac_dev * dev,u32 size)619 static int aac_src_ioremap(struct aac_dev *dev, u32 size)
620 {
621 if (!size) {
622 iounmap(dev->regs.src.bar1);
623 dev->regs.src.bar1 = NULL;
624 iounmap(dev->regs.src.bar0);
625 dev->base = dev->regs.src.bar0 = NULL;
626 return 0;
627 }
628 dev->regs.src.bar1 = ioremap(pci_resource_start(dev->pdev, 2),
629 AAC_MIN_SRC_BAR1_SIZE);
630 dev->base = NULL;
631 if (dev->regs.src.bar1 == NULL)
632 return -1;
633 dev->base = dev->regs.src.bar0 = ioremap(dev->base_start, size);
634 if (dev->base == NULL) {
635 iounmap(dev->regs.src.bar1);
636 dev->regs.src.bar1 = NULL;
637 return -1;
638 }
639 dev->IndexRegs = &((struct src_registers __iomem *)
640 dev->base)->u.tupelo.IndexRegs;
641 return 0;
642 }
643
644 /**
645 * aac_srcv_ioremap
646 * @size: mapping resize request
647 *
648 */
aac_srcv_ioremap(struct aac_dev * dev,u32 size)649 static int aac_srcv_ioremap(struct aac_dev *dev, u32 size)
650 {
651 if (!size) {
652 iounmap(dev->regs.src.bar0);
653 dev->base = dev->regs.src.bar0 = NULL;
654 return 0;
655 }
656
657 dev->regs.src.bar1 =
658 ioremap(pci_resource_start(dev->pdev, 2), AAC_MIN_SRCV_BAR1_SIZE);
659 dev->base = NULL;
660 if (dev->regs.src.bar1 == NULL)
661 return -1;
662 dev->base = dev->regs.src.bar0 = ioremap(dev->base_start, size);
663 if (dev->base == NULL) {
664 iounmap(dev->regs.src.bar1);
665 dev->regs.src.bar1 = NULL;
666 return -1;
667 }
668 dev->IndexRegs = &((struct src_registers __iomem *)
669 dev->base)->u.denali.IndexRegs;
670 return 0;
671 }
672
aac_set_intx_mode(struct aac_dev * dev)673 void aac_set_intx_mode(struct aac_dev *dev)
674 {
675 if (dev->msi_enabled) {
676 aac_src_access_devreg(dev, AAC_ENABLE_INTX);
677 dev->msi_enabled = 0;
678 msleep(5000); /* Delay 5 seconds */
679 }
680 }
681
aac_dump_fw_fib_iop_reset(struct aac_dev * dev)682 static void aac_dump_fw_fib_iop_reset(struct aac_dev *dev)
683 {
684 __le32 supported_options3;
685
686 if (!aac_fib_dump)
687 return;
688
689 supported_options3 = dev->supplement_adapter_info.supported_options3;
690 if (!(supported_options3 & AAC_OPTION_SUPPORTED3_IOP_RESET_FIB_DUMP))
691 return;
692
693 aac_adapter_sync_cmd(dev, IOP_RESET_FW_FIB_DUMP,
694 0, 0, 0, 0, 0, 0, NULL, NULL, NULL, NULL, NULL);
695 }
696
aac_is_ctrl_up_and_running(struct aac_dev * dev)697 static bool aac_is_ctrl_up_and_running(struct aac_dev *dev)
698 {
699 bool ctrl_up = true;
700 unsigned long status, start;
701 bool is_up = false;
702
703 start = jiffies;
704 do {
705 schedule();
706 status = src_readl(dev, MUnit.OMR);
707
708 if (status == 0xffffffff)
709 status = 0;
710
711 if (status & KERNEL_BOOTING) {
712 start = jiffies;
713 continue;
714 }
715
716 if (time_after(jiffies, start+HZ*SOFT_RESET_TIME)) {
717 ctrl_up = false;
718 break;
719 }
720
721 is_up = status & KERNEL_UP_AND_RUNNING;
722
723 } while (!is_up);
724
725 return ctrl_up;
726 }
727
aac_notify_fw_of_iop_reset(struct aac_dev * dev)728 static void aac_notify_fw_of_iop_reset(struct aac_dev *dev)
729 {
730 aac_adapter_sync_cmd(dev, IOP_RESET_ALWAYS, 0, 0, 0, 0, 0, 0, NULL,
731 NULL, NULL, NULL, NULL);
732 }
733
aac_send_iop_reset(struct aac_dev * dev)734 static void aac_send_iop_reset(struct aac_dev *dev)
735 {
736 aac_dump_fw_fib_iop_reset(dev);
737
738 aac_notify_fw_of_iop_reset(dev);
739
740 aac_set_intx_mode(dev);
741
742 src_writel(dev, MUnit.IDR, IOP_SRC_RESET_MASK);
743
744 msleep(5000);
745 }
746
aac_send_hardware_soft_reset(struct aac_dev * dev)747 static void aac_send_hardware_soft_reset(struct aac_dev *dev)
748 {
749 u_int32_t val;
750
751 val = readl(((char *)(dev->base) + IBW_SWR_OFFSET));
752 val |= 0x01;
753 writel(val, ((char *)(dev->base) + IBW_SWR_OFFSET));
754 msleep_interruptible(20000);
755 }
756
aac_src_restart_adapter(struct aac_dev * dev,int bled,u8 reset_type)757 static int aac_src_restart_adapter(struct aac_dev *dev, int bled, u8 reset_type)
758 {
759 bool is_ctrl_up;
760 int ret = 0;
761
762 if (bled < 0)
763 goto invalid_out;
764
765 if (bled)
766 dev_err(&dev->pdev->dev, "adapter kernel panic'd %x.\n", bled);
767
768 /*
769 * When there is a BlinkLED, IOP_RESET has not effect
770 */
771 if (bled >= 2 && dev->sa_firmware && reset_type & HW_IOP_RESET)
772 reset_type &= ~HW_IOP_RESET;
773
774 dev->a_ops.adapter_enable_int = aac_src_disable_interrupt;
775
776 dev_err(&dev->pdev->dev, "Controller reset type is %d\n", reset_type);
777
778 if (reset_type & HW_IOP_RESET) {
779 dev_info(&dev->pdev->dev, "Issuing IOP reset\n");
780 aac_send_iop_reset(dev);
781
782 /*
783 * Creates a delay or wait till up and running comes thru
784 */
785 is_ctrl_up = aac_is_ctrl_up_and_running(dev);
786 if (!is_ctrl_up)
787 dev_err(&dev->pdev->dev, "IOP reset failed\n");
788 else {
789 dev_info(&dev->pdev->dev, "IOP reset succeded\n");
790 goto set_startup;
791 }
792 }
793
794 if (!dev->sa_firmware) {
795 dev_err(&dev->pdev->dev, "ARC Reset attempt failed\n");
796 ret = -ENODEV;
797 goto out;
798 }
799
800 if (reset_type & HW_SOFT_RESET) {
801 dev_info(&dev->pdev->dev, "Issuing SOFT reset\n");
802 aac_send_hardware_soft_reset(dev);
803 dev->msi_enabled = 0;
804
805 is_ctrl_up = aac_is_ctrl_up_and_running(dev);
806 if (!is_ctrl_up) {
807 dev_err(&dev->pdev->dev, "SOFT reset failed\n");
808 ret = -ENODEV;
809 goto out;
810 } else
811 dev_info(&dev->pdev->dev, "SOFT reset succeded\n");
812 }
813
814 set_startup:
815 if (startup_timeout < 300)
816 startup_timeout = 300;
817
818 out:
819 return ret;
820
821 invalid_out:
822 if (src_readl(dev, MUnit.OMR) & KERNEL_PANIC)
823 ret = -ENODEV;
824 goto out;
825 }
826
827 /**
828 * aac_src_select_comm - Select communications method
829 * @dev: Adapter
830 * @comm: communications method
831 */
aac_src_select_comm(struct aac_dev * dev,int comm)832 static int aac_src_select_comm(struct aac_dev *dev, int comm)
833 {
834 switch (comm) {
835 case AAC_COMM_MESSAGE:
836 dev->a_ops.adapter_intr = aac_src_intr_message;
837 dev->a_ops.adapter_deliver = aac_src_deliver_message;
838 break;
839 default:
840 return 1;
841 }
842 return 0;
843 }
844
845 /**
846 * aac_src_init - initialize an Cardinal Frey Bar card
847 * @dev: device to configure
848 *
849 */
850
aac_src_init(struct aac_dev * dev)851 int aac_src_init(struct aac_dev *dev)
852 {
853 unsigned long start;
854 unsigned long status;
855 int restart = 0;
856 int instance = dev->id;
857 const char *name = dev->name;
858
859 dev->a_ops.adapter_ioremap = aac_src_ioremap;
860 dev->a_ops.adapter_comm = aac_src_select_comm;
861
862 dev->base_size = AAC_MIN_SRC_BAR0_SIZE;
863 if (aac_adapter_ioremap(dev, dev->base_size)) {
864 printk(KERN_WARNING "%s: unable to map adapter.\n", name);
865 goto error_iounmap;
866 }
867
868 /* Failure to reset here is an option ... */
869 dev->a_ops.adapter_sync_cmd = src_sync_cmd;
870 dev->a_ops.adapter_enable_int = aac_src_disable_interrupt;
871 if ((aac_reset_devices || reset_devices) &&
872 !aac_src_restart_adapter(dev, 0, IOP_HWSOFT_RESET))
873 ++restart;
874 /*
875 * Check to see if the board panic'd while booting.
876 */
877 status = src_readl(dev, MUnit.OMR);
878 if (status & KERNEL_PANIC) {
879 if (aac_src_restart_adapter(dev,
880 aac_src_check_health(dev), IOP_HWSOFT_RESET))
881 goto error_iounmap;
882 ++restart;
883 }
884 /*
885 * Check to see if the board failed any self tests.
886 */
887 status = src_readl(dev, MUnit.OMR);
888 if (status & SELF_TEST_FAILED) {
889 printk(KERN_ERR "%s%d: adapter self-test failed.\n",
890 dev->name, instance);
891 goto error_iounmap;
892 }
893 /*
894 * Check to see if the monitor panic'd while booting.
895 */
896 if (status & MONITOR_PANIC) {
897 printk(KERN_ERR "%s%d: adapter monitor panic.\n",
898 dev->name, instance);
899 goto error_iounmap;
900 }
901 start = jiffies;
902 /*
903 * Wait for the adapter to be up and running. Wait up to 3 minutes
904 */
905 while (!((status = src_readl(dev, MUnit.OMR)) &
906 KERNEL_UP_AND_RUNNING)) {
907 if ((restart &&
908 (status & (KERNEL_PANIC|SELF_TEST_FAILED|MONITOR_PANIC))) ||
909 time_after(jiffies, start+HZ*startup_timeout)) {
910 printk(KERN_ERR "%s%d: adapter kernel failed to start, init status = %lx.\n",
911 dev->name, instance, status);
912 goto error_iounmap;
913 }
914 if (!restart &&
915 ((status & (KERNEL_PANIC|SELF_TEST_FAILED|MONITOR_PANIC)) ||
916 time_after(jiffies, start + HZ *
917 ((startup_timeout > 60)
918 ? (startup_timeout - 60)
919 : (startup_timeout / 2))))) {
920 if (likely(!aac_src_restart_adapter(dev,
921 aac_src_check_health(dev), IOP_HWSOFT_RESET)))
922 start = jiffies;
923 ++restart;
924 }
925 msleep(1);
926 }
927 if (restart && aac_commit)
928 aac_commit = 1;
929 /*
930 * Fill in the common function dispatch table.
931 */
932 dev->a_ops.adapter_interrupt = aac_src_interrupt_adapter;
933 dev->a_ops.adapter_disable_int = aac_src_disable_interrupt;
934 dev->a_ops.adapter_enable_int = aac_src_disable_interrupt;
935 dev->a_ops.adapter_notify = aac_src_notify_adapter;
936 dev->a_ops.adapter_sync_cmd = src_sync_cmd;
937 dev->a_ops.adapter_check_health = aac_src_check_health;
938 dev->a_ops.adapter_restart = aac_src_restart_adapter;
939 dev->a_ops.adapter_start = aac_src_start_adapter;
940
941 /*
942 * First clear out all interrupts. Then enable the one's that we
943 * can handle.
944 */
945 aac_adapter_comm(dev, AAC_COMM_MESSAGE);
946 aac_adapter_disable_int(dev);
947 src_writel(dev, MUnit.ODR_C, 0xffffffff);
948 aac_adapter_enable_int(dev);
949
950 if (aac_init_adapter(dev) == NULL)
951 goto error_iounmap;
952 if (dev->comm_interface != AAC_COMM_MESSAGE_TYPE1)
953 goto error_iounmap;
954
955 dev->msi = !pci_enable_msi(dev->pdev);
956
957 dev->aac_msix[0].vector_no = 0;
958 dev->aac_msix[0].dev = dev;
959
960 if (request_irq(dev->pdev->irq, dev->a_ops.adapter_intr,
961 IRQF_SHARED, "aacraid", &(dev->aac_msix[0])) < 0) {
962
963 if (dev->msi)
964 pci_disable_msi(dev->pdev);
965
966 printk(KERN_ERR "%s%d: Interrupt unavailable.\n",
967 name, instance);
968 goto error_iounmap;
969 }
970 dev->dbg_base = pci_resource_start(dev->pdev, 2);
971 dev->dbg_base_mapped = dev->regs.src.bar1;
972 dev->dbg_size = AAC_MIN_SRC_BAR1_SIZE;
973 dev->a_ops.adapter_enable_int = aac_src_enable_interrupt_message;
974
975 aac_adapter_enable_int(dev);
976
977 if (!dev->sync_mode) {
978 /*
979 * Tell the adapter that all is configured, and it can
980 * start accepting requests
981 */
982 aac_src_start_adapter(dev);
983 }
984 return 0;
985
986 error_iounmap:
987
988 return -1;
989 }
990
991 /**
992 * aac_srcv_init - initialize an SRCv card
993 * @dev: device to configure
994 *
995 */
996
aac_srcv_init(struct aac_dev * dev)997 int aac_srcv_init(struct aac_dev *dev)
998 {
999 unsigned long start;
1000 unsigned long status;
1001 int restart = 0;
1002 int instance = dev->id;
1003 const char *name = dev->name;
1004
1005 dev->a_ops.adapter_ioremap = aac_srcv_ioremap;
1006 dev->a_ops.adapter_comm = aac_src_select_comm;
1007
1008 dev->base_size = AAC_MIN_SRCV_BAR0_SIZE;
1009 if (aac_adapter_ioremap(dev, dev->base_size)) {
1010 printk(KERN_WARNING "%s: unable to map adapter.\n", name);
1011 goto error_iounmap;
1012 }
1013
1014 /* Failure to reset here is an option ... */
1015 dev->a_ops.adapter_sync_cmd = src_sync_cmd;
1016 dev->a_ops.adapter_enable_int = aac_src_disable_interrupt;
1017 if ((aac_reset_devices || reset_devices) &&
1018 !aac_src_restart_adapter(dev, 0, IOP_HWSOFT_RESET))
1019 ++restart;
1020 /*
1021 * Check to see if flash update is running.
1022 * Wait for the adapter to be up and running. Wait up to 5 minutes
1023 */
1024 status = src_readl(dev, MUnit.OMR);
1025 if (status & FLASH_UPD_PENDING) {
1026 start = jiffies;
1027 do {
1028 status = src_readl(dev, MUnit.OMR);
1029 if (time_after(jiffies, start+HZ*FWUPD_TIMEOUT)) {
1030 printk(KERN_ERR "%s%d: adapter flash update failed.\n",
1031 dev->name, instance);
1032 goto error_iounmap;
1033 }
1034 } while (!(status & FLASH_UPD_SUCCESS) &&
1035 !(status & FLASH_UPD_FAILED));
1036 /* Delay 10 seconds.
1037 * Because right now FW is doing a soft reset,
1038 * do not read scratch pad register at this time
1039 */
1040 ssleep(10);
1041 }
1042 /*
1043 * Check to see if the board panic'd while booting.
1044 */
1045 status = src_readl(dev, MUnit.OMR);
1046 if (status & KERNEL_PANIC) {
1047 if (aac_src_restart_adapter(dev,
1048 aac_src_check_health(dev), IOP_HWSOFT_RESET))
1049 goto error_iounmap;
1050 ++restart;
1051 }
1052 /*
1053 * Check to see if the board failed any self tests.
1054 */
1055 status = src_readl(dev, MUnit.OMR);
1056 if (status & SELF_TEST_FAILED) {
1057 printk(KERN_ERR "%s%d: adapter self-test failed.\n", dev->name, instance);
1058 goto error_iounmap;
1059 }
1060 /*
1061 * Check to see if the monitor panic'd while booting.
1062 */
1063 if (status & MONITOR_PANIC) {
1064 printk(KERN_ERR "%s%d: adapter monitor panic.\n", dev->name, instance);
1065 goto error_iounmap;
1066 }
1067 start = jiffies;
1068 /*
1069 * Wait for the adapter to be up and running. Wait up to 3 minutes
1070 */
1071 while (!((status = src_readl(dev, MUnit.OMR)) &
1072 KERNEL_UP_AND_RUNNING) ||
1073 status == 0xffffffff) {
1074 if ((restart &&
1075 (status & (KERNEL_PANIC|SELF_TEST_FAILED|MONITOR_PANIC))) ||
1076 time_after(jiffies, start+HZ*startup_timeout)) {
1077 printk(KERN_ERR "%s%d: adapter kernel failed to start, init status = %lx.\n",
1078 dev->name, instance, status);
1079 goto error_iounmap;
1080 }
1081 if (!restart &&
1082 ((status & (KERNEL_PANIC|SELF_TEST_FAILED|MONITOR_PANIC)) ||
1083 time_after(jiffies, start + HZ *
1084 ((startup_timeout > 60)
1085 ? (startup_timeout - 60)
1086 : (startup_timeout / 2))))) {
1087 if (likely(!aac_src_restart_adapter(dev,
1088 aac_src_check_health(dev), IOP_HWSOFT_RESET)))
1089 start = jiffies;
1090 ++restart;
1091 }
1092 msleep(1);
1093 }
1094 if (restart && aac_commit)
1095 aac_commit = 1;
1096 /*
1097 * Fill in the common function dispatch table.
1098 */
1099 dev->a_ops.adapter_interrupt = aac_src_interrupt_adapter;
1100 dev->a_ops.adapter_disable_int = aac_src_disable_interrupt;
1101 dev->a_ops.adapter_enable_int = aac_src_disable_interrupt;
1102 dev->a_ops.adapter_notify = aac_src_notify_adapter;
1103 dev->a_ops.adapter_sync_cmd = src_sync_cmd;
1104 dev->a_ops.adapter_check_health = aac_src_check_health;
1105 dev->a_ops.adapter_restart = aac_src_restart_adapter;
1106 dev->a_ops.adapter_start = aac_src_start_adapter;
1107
1108 /*
1109 * First clear out all interrupts. Then enable the one's that we
1110 * can handle.
1111 */
1112 aac_adapter_comm(dev, AAC_COMM_MESSAGE);
1113 aac_adapter_disable_int(dev);
1114 src_writel(dev, MUnit.ODR_C, 0xffffffff);
1115 aac_adapter_enable_int(dev);
1116
1117 if (aac_init_adapter(dev) == NULL)
1118 goto error_iounmap;
1119 if ((dev->comm_interface != AAC_COMM_MESSAGE_TYPE2) &&
1120 (dev->comm_interface != AAC_COMM_MESSAGE_TYPE3))
1121 goto error_iounmap;
1122 if (dev->msi_enabled)
1123 aac_src_access_devreg(dev, AAC_ENABLE_MSIX);
1124
1125 if (aac_acquire_irq(dev))
1126 goto error_iounmap;
1127
1128 dev->dbg_base = pci_resource_start(dev->pdev, 2);
1129 dev->dbg_base_mapped = dev->regs.src.bar1;
1130 dev->dbg_size = AAC_MIN_SRCV_BAR1_SIZE;
1131 dev->a_ops.adapter_enable_int = aac_src_enable_interrupt_message;
1132
1133 aac_adapter_enable_int(dev);
1134
1135 if (!dev->sync_mode) {
1136 /*
1137 * Tell the adapter that all is configured, and it can
1138 * start accepting requests
1139 */
1140 aac_src_start_adapter(dev);
1141 }
1142 return 0;
1143
1144 error_iounmap:
1145
1146 return -1;
1147 }
1148
aac_src_access_devreg(struct aac_dev * dev,int mode)1149 void aac_src_access_devreg(struct aac_dev *dev, int mode)
1150 {
1151 u_int32_t val;
1152
1153 switch (mode) {
1154 case AAC_ENABLE_INTERRUPT:
1155 src_writel(dev,
1156 MUnit.OIMR,
1157 dev->OIMR = (dev->msi_enabled ?
1158 AAC_INT_ENABLE_TYPE1_MSIX :
1159 AAC_INT_ENABLE_TYPE1_INTX));
1160 break;
1161
1162 case AAC_DISABLE_INTERRUPT:
1163 src_writel(dev,
1164 MUnit.OIMR,
1165 dev->OIMR = AAC_INT_DISABLE_ALL);
1166 break;
1167
1168 case AAC_ENABLE_MSIX:
1169 /* set bit 6 */
1170 val = src_readl(dev, MUnit.IDR);
1171 val |= 0x40;
1172 src_writel(dev, MUnit.IDR, val);
1173 src_readl(dev, MUnit.IDR);
1174 /* unmask int. */
1175 val = PMC_ALL_INTERRUPT_BITS;
1176 src_writel(dev, MUnit.IOAR, val);
1177 val = src_readl(dev, MUnit.OIMR);
1178 src_writel(dev,
1179 MUnit.OIMR,
1180 val & (~(PMC_GLOBAL_INT_BIT2 | PMC_GLOBAL_INT_BIT0)));
1181 break;
1182
1183 case AAC_DISABLE_MSIX:
1184 /* reset bit 6 */
1185 val = src_readl(dev, MUnit.IDR);
1186 val &= ~0x40;
1187 src_writel(dev, MUnit.IDR, val);
1188 src_readl(dev, MUnit.IDR);
1189 break;
1190
1191 case AAC_CLEAR_AIF_BIT:
1192 /* set bit 5 */
1193 val = src_readl(dev, MUnit.IDR);
1194 val |= 0x20;
1195 src_writel(dev, MUnit.IDR, val);
1196 src_readl(dev, MUnit.IDR);
1197 break;
1198
1199 case AAC_CLEAR_SYNC_BIT:
1200 /* set bit 4 */
1201 val = src_readl(dev, MUnit.IDR);
1202 val |= 0x10;
1203 src_writel(dev, MUnit.IDR, val);
1204 src_readl(dev, MUnit.IDR);
1205 break;
1206
1207 case AAC_ENABLE_INTX:
1208 /* set bit 7 */
1209 val = src_readl(dev, MUnit.IDR);
1210 val |= 0x80;
1211 src_writel(dev, MUnit.IDR, val);
1212 src_readl(dev, MUnit.IDR);
1213 /* unmask int. */
1214 val = PMC_ALL_INTERRUPT_BITS;
1215 src_writel(dev, MUnit.IOAR, val);
1216 src_readl(dev, MUnit.IOAR);
1217 val = src_readl(dev, MUnit.OIMR);
1218 src_writel(dev, MUnit.OIMR,
1219 val & (~(PMC_GLOBAL_INT_BIT2)));
1220 break;
1221
1222 default:
1223 break;
1224 }
1225 }
1226
aac_src_get_sync_status(struct aac_dev * dev)1227 static int aac_src_get_sync_status(struct aac_dev *dev)
1228 {
1229
1230 int val;
1231
1232 if (dev->msi_enabled)
1233 val = src_readl(dev, MUnit.ODR_MSI) & 0x1000 ? 1 : 0;
1234 else
1235 val = src_readl(dev, MUnit.ODR_R) >> SRC_ODR_SHIFT;
1236
1237 return val;
1238 }
1239