1 /*
2 * i2c-xiic.c
3 * Copyright (c) 2002-2007 Xilinx Inc.
4 * Copyright (c) 2009-2010 Intel Corporation
5 *
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License version 2 as
8 * published by the Free Software Foundation.
9 *
10 * This program is distributed in the hope that it will be useful,
11 * but WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
13 * GNU General Public License for more details.
14 *
15 *
16 * This code was implemented by Mocean Laboratories AB when porting linux
17 * to the automotive development board Russellville. The copyright holder
18 * as seen in the header is Intel corporation.
19 * Mocean Laboratories forked off the GNU/Linux platform work into a
20 * separate company called Pelagicore AB, which committed the code to the
21 * kernel.
22 */
23
24 /* Supports:
25 * Xilinx IIC
26 */
27 #include <linux/kernel.h>
28 #include <linux/module.h>
29 #include <linux/errno.h>
30 #include <linux/err.h>
31 #include <linux/delay.h>
32 #include <linux/platform_device.h>
33 #include <linux/i2c.h>
34 #include <linux/interrupt.h>
35 #include <linux/wait.h>
36 #include <linux/i2c-xiic.h>
37 #include <linux/io.h>
38 #include <linux/slab.h>
39 #include <linux/of.h>
40
41 #define DRIVER_NAME "xiic-i2c"
42
43 enum xilinx_i2c_state {
44 STATE_DONE,
45 STATE_ERROR,
46 STATE_START
47 };
48
49 enum xiic_endian {
50 LITTLE,
51 BIG
52 };
53
54 /**
55 * struct xiic_i2c - Internal representation of the XIIC I2C bus
56 * @base: Memory base of the HW registers
57 * @wait: Wait queue for callers
58 * @adap: Kernel adapter representation
59 * @tx_msg: Messages from above to be sent
60 * @lock: Mutual exclusion
61 * @tx_pos: Current pos in TX message
62 * @nmsgs: Number of messages in tx_msg
63 * @state: See STATE_
64 * @rx_msg: Current RX message
65 * @rx_pos: Position within current RX message
66 * @endianness: big/little-endian byte order
67 */
68 struct xiic_i2c {
69 void __iomem *base;
70 wait_queue_head_t wait;
71 struct i2c_adapter adap;
72 struct i2c_msg *tx_msg;
73 spinlock_t lock;
74 unsigned int tx_pos;
75 unsigned int nmsgs;
76 enum xilinx_i2c_state state;
77 struct i2c_msg *rx_msg;
78 int rx_pos;
79 enum xiic_endian endianness;
80 };
81
82
83 #define XIIC_MSB_OFFSET 0
84 #define XIIC_REG_OFFSET (0x100+XIIC_MSB_OFFSET)
85
86 /*
87 * Register offsets in bytes from RegisterBase. Three is added to the
88 * base offset to access LSB (IBM style) of the word
89 */
90 #define XIIC_CR_REG_OFFSET (0x00+XIIC_REG_OFFSET) /* Control Register */
91 #define XIIC_SR_REG_OFFSET (0x04+XIIC_REG_OFFSET) /* Status Register */
92 #define XIIC_DTR_REG_OFFSET (0x08+XIIC_REG_OFFSET) /* Data Tx Register */
93 #define XIIC_DRR_REG_OFFSET (0x0C+XIIC_REG_OFFSET) /* Data Rx Register */
94 #define XIIC_ADR_REG_OFFSET (0x10+XIIC_REG_OFFSET) /* Address Register */
95 #define XIIC_TFO_REG_OFFSET (0x14+XIIC_REG_OFFSET) /* Tx FIFO Occupancy */
96 #define XIIC_RFO_REG_OFFSET (0x18+XIIC_REG_OFFSET) /* Rx FIFO Occupancy */
97 #define XIIC_TBA_REG_OFFSET (0x1C+XIIC_REG_OFFSET) /* 10 Bit Address reg */
98 #define XIIC_RFD_REG_OFFSET (0x20+XIIC_REG_OFFSET) /* Rx FIFO Depth reg */
99 #define XIIC_GPO_REG_OFFSET (0x24+XIIC_REG_OFFSET) /* Output Register */
100
101 /* Control Register masks */
102 #define XIIC_CR_ENABLE_DEVICE_MASK 0x01 /* Device enable = 1 */
103 #define XIIC_CR_TX_FIFO_RESET_MASK 0x02 /* Transmit FIFO reset=1 */
104 #define XIIC_CR_MSMS_MASK 0x04 /* Master starts Txing=1 */
105 #define XIIC_CR_DIR_IS_TX_MASK 0x08 /* Dir of tx. Txing=1 */
106 #define XIIC_CR_NO_ACK_MASK 0x10 /* Tx Ack. NO ack = 1 */
107 #define XIIC_CR_REPEATED_START_MASK 0x20 /* Repeated start = 1 */
108 #define XIIC_CR_GENERAL_CALL_MASK 0x40 /* Gen Call enabled = 1 */
109
110 /* Status Register masks */
111 #define XIIC_SR_GEN_CALL_MASK 0x01 /* 1=a mstr issued a GC */
112 #define XIIC_SR_ADDR_AS_SLAVE_MASK 0x02 /* 1=when addr as slave */
113 #define XIIC_SR_BUS_BUSY_MASK 0x04 /* 1 = bus is busy */
114 #define XIIC_SR_MSTR_RDING_SLAVE_MASK 0x08 /* 1=Dir: mstr <-- slave */
115 #define XIIC_SR_TX_FIFO_FULL_MASK 0x10 /* 1 = Tx FIFO full */
116 #define XIIC_SR_RX_FIFO_FULL_MASK 0x20 /* 1 = Rx FIFO full */
117 #define XIIC_SR_RX_FIFO_EMPTY_MASK 0x40 /* 1 = Rx FIFO empty */
118 #define XIIC_SR_TX_FIFO_EMPTY_MASK 0x80 /* 1 = Tx FIFO empty */
119
120 /* Interrupt Status Register masks Interrupt occurs when... */
121 #define XIIC_INTR_ARB_LOST_MASK 0x01 /* 1 = arbitration lost */
122 #define XIIC_INTR_TX_ERROR_MASK 0x02 /* 1=Tx error/msg complete */
123 #define XIIC_INTR_TX_EMPTY_MASK 0x04 /* 1 = Tx FIFO/reg empty */
124 #define XIIC_INTR_RX_FULL_MASK 0x08 /* 1=Rx FIFO/reg=OCY level */
125 #define XIIC_INTR_BNB_MASK 0x10 /* 1 = Bus not busy */
126 #define XIIC_INTR_AAS_MASK 0x20 /* 1 = when addr as slave */
127 #define XIIC_INTR_NAAS_MASK 0x40 /* 1 = not addr as slave */
128 #define XIIC_INTR_TX_HALF_MASK 0x80 /* 1 = TX FIFO half empty */
129
130 /* The following constants specify the depth of the FIFOs */
131 #define IIC_RX_FIFO_DEPTH 16 /* Rx fifo capacity */
132 #define IIC_TX_FIFO_DEPTH 16 /* Tx fifo capacity */
133
134 /* The following constants specify groups of interrupts that are typically
135 * enabled or disables at the same time
136 */
137 #define XIIC_TX_INTERRUPTS \
138 (XIIC_INTR_TX_ERROR_MASK | XIIC_INTR_TX_EMPTY_MASK | XIIC_INTR_TX_HALF_MASK)
139
140 #define XIIC_TX_RX_INTERRUPTS (XIIC_INTR_RX_FULL_MASK | XIIC_TX_INTERRUPTS)
141
142 /* The following constants are used with the following macros to specify the
143 * operation, a read or write operation.
144 */
145 #define XIIC_READ_OPERATION 1
146 #define XIIC_WRITE_OPERATION 0
147
148 /*
149 * Tx Fifo upper bit masks.
150 */
151 #define XIIC_TX_DYN_START_MASK 0x0100 /* 1 = Set dynamic start */
152 #define XIIC_TX_DYN_STOP_MASK 0x0200 /* 1 = Set dynamic stop */
153
154 /*
155 * The following constants define the register offsets for the Interrupt
156 * registers. There are some holes in the memory map for reserved addresses
157 * to allow other registers to be added and still match the memory map of the
158 * interrupt controller registers
159 */
160 #define XIIC_DGIER_OFFSET 0x1C /* Device Global Interrupt Enable Register */
161 #define XIIC_IISR_OFFSET 0x20 /* Interrupt Status Register */
162 #define XIIC_IIER_OFFSET 0x28 /* Interrupt Enable Register */
163 #define XIIC_RESETR_OFFSET 0x40 /* Reset Register */
164
165 #define XIIC_RESET_MASK 0xAUL
166
167 /*
168 * The following constant is used for the device global interrupt enable
169 * register, to enable all interrupts for the device, this is the only bit
170 * in the register
171 */
172 #define XIIC_GINTR_ENABLE_MASK 0x80000000UL
173
174 #define xiic_tx_space(i2c) ((i2c)->tx_msg->len - (i2c)->tx_pos)
175 #define xiic_rx_space(i2c) ((i2c)->rx_msg->len - (i2c)->rx_pos)
176
177 static void xiic_start_xfer(struct xiic_i2c *i2c);
178 static void __xiic_start_xfer(struct xiic_i2c *i2c);
179
180 /*
181 * For the register read and write functions, a little-endian and big-endian
182 * version are necessary. Endianness is detected during the probe function.
183 * Only the least significant byte [doublet] of the register are ever
184 * accessed. This requires an offset of 3 [2] from the base address for
185 * big-endian systems.
186 */
187
xiic_setreg8(struct xiic_i2c * i2c,int reg,u8 value)188 static inline void xiic_setreg8(struct xiic_i2c *i2c, int reg, u8 value)
189 {
190 if (i2c->endianness == LITTLE)
191 iowrite8(value, i2c->base + reg);
192 else
193 iowrite8(value, i2c->base + reg + 3);
194 }
195
xiic_getreg8(struct xiic_i2c * i2c,int reg)196 static inline u8 xiic_getreg8(struct xiic_i2c *i2c, int reg)
197 {
198 u8 ret;
199
200 if (i2c->endianness == LITTLE)
201 ret = ioread8(i2c->base + reg);
202 else
203 ret = ioread8(i2c->base + reg + 3);
204 return ret;
205 }
206
xiic_setreg16(struct xiic_i2c * i2c,int reg,u16 value)207 static inline void xiic_setreg16(struct xiic_i2c *i2c, int reg, u16 value)
208 {
209 if (i2c->endianness == LITTLE)
210 iowrite16(value, i2c->base + reg);
211 else
212 iowrite16be(value, i2c->base + reg + 2);
213 }
214
xiic_setreg32(struct xiic_i2c * i2c,int reg,int value)215 static inline void xiic_setreg32(struct xiic_i2c *i2c, int reg, int value)
216 {
217 if (i2c->endianness == LITTLE)
218 iowrite32(value, i2c->base + reg);
219 else
220 iowrite32be(value, i2c->base + reg);
221 }
222
xiic_getreg32(struct xiic_i2c * i2c,int reg)223 static inline int xiic_getreg32(struct xiic_i2c *i2c, int reg)
224 {
225 u32 ret;
226
227 if (i2c->endianness == LITTLE)
228 ret = ioread32(i2c->base + reg);
229 else
230 ret = ioread32be(i2c->base + reg);
231 return ret;
232 }
233
xiic_irq_dis(struct xiic_i2c * i2c,u32 mask)234 static inline void xiic_irq_dis(struct xiic_i2c *i2c, u32 mask)
235 {
236 u32 ier = xiic_getreg32(i2c, XIIC_IIER_OFFSET);
237 xiic_setreg32(i2c, XIIC_IIER_OFFSET, ier & ~mask);
238 }
239
xiic_irq_en(struct xiic_i2c * i2c,u32 mask)240 static inline void xiic_irq_en(struct xiic_i2c *i2c, u32 mask)
241 {
242 u32 ier = xiic_getreg32(i2c, XIIC_IIER_OFFSET);
243 xiic_setreg32(i2c, XIIC_IIER_OFFSET, ier | mask);
244 }
245
xiic_irq_clr(struct xiic_i2c * i2c,u32 mask)246 static inline void xiic_irq_clr(struct xiic_i2c *i2c, u32 mask)
247 {
248 u32 isr = xiic_getreg32(i2c, XIIC_IISR_OFFSET);
249 xiic_setreg32(i2c, XIIC_IISR_OFFSET, isr & mask);
250 }
251
xiic_irq_clr_en(struct xiic_i2c * i2c,u32 mask)252 static inline void xiic_irq_clr_en(struct xiic_i2c *i2c, u32 mask)
253 {
254 xiic_irq_clr(i2c, mask);
255 xiic_irq_en(i2c, mask);
256 }
257
xiic_clear_rx_fifo(struct xiic_i2c * i2c)258 static void xiic_clear_rx_fifo(struct xiic_i2c *i2c)
259 {
260 u8 sr;
261 for (sr = xiic_getreg8(i2c, XIIC_SR_REG_OFFSET);
262 !(sr & XIIC_SR_RX_FIFO_EMPTY_MASK);
263 sr = xiic_getreg8(i2c, XIIC_SR_REG_OFFSET))
264 xiic_getreg8(i2c, XIIC_DRR_REG_OFFSET);
265 }
266
xiic_reinit(struct xiic_i2c * i2c)267 static void xiic_reinit(struct xiic_i2c *i2c)
268 {
269 xiic_setreg32(i2c, XIIC_RESETR_OFFSET, XIIC_RESET_MASK);
270
271 /* Set receive Fifo depth to maximum (zero based). */
272 xiic_setreg8(i2c, XIIC_RFD_REG_OFFSET, IIC_RX_FIFO_DEPTH - 1);
273
274 /* Reset Tx Fifo. */
275 xiic_setreg8(i2c, XIIC_CR_REG_OFFSET, XIIC_CR_TX_FIFO_RESET_MASK);
276
277 /* Enable IIC Device, remove Tx Fifo reset & disable general call. */
278 xiic_setreg8(i2c, XIIC_CR_REG_OFFSET, XIIC_CR_ENABLE_DEVICE_MASK);
279
280 /* make sure RX fifo is empty */
281 xiic_clear_rx_fifo(i2c);
282
283 /* Enable interrupts */
284 xiic_setreg32(i2c, XIIC_DGIER_OFFSET, XIIC_GINTR_ENABLE_MASK);
285
286 xiic_irq_clr_en(i2c, XIIC_INTR_ARB_LOST_MASK);
287 }
288
xiic_deinit(struct xiic_i2c * i2c)289 static void xiic_deinit(struct xiic_i2c *i2c)
290 {
291 u8 cr;
292
293 xiic_setreg32(i2c, XIIC_RESETR_OFFSET, XIIC_RESET_MASK);
294
295 /* Disable IIC Device. */
296 cr = xiic_getreg8(i2c, XIIC_CR_REG_OFFSET);
297 xiic_setreg8(i2c, XIIC_CR_REG_OFFSET, cr & ~XIIC_CR_ENABLE_DEVICE_MASK);
298 }
299
xiic_read_rx(struct xiic_i2c * i2c)300 static void xiic_read_rx(struct xiic_i2c *i2c)
301 {
302 u8 bytes_in_fifo;
303 int i;
304
305 bytes_in_fifo = xiic_getreg8(i2c, XIIC_RFO_REG_OFFSET) + 1;
306
307 dev_dbg(i2c->adap.dev.parent,
308 "%s entry, bytes in fifo: %d, msg: %d, SR: 0x%x, CR: 0x%x\n",
309 __func__, bytes_in_fifo, xiic_rx_space(i2c),
310 xiic_getreg8(i2c, XIIC_SR_REG_OFFSET),
311 xiic_getreg8(i2c, XIIC_CR_REG_OFFSET));
312
313 if (bytes_in_fifo > xiic_rx_space(i2c))
314 bytes_in_fifo = xiic_rx_space(i2c);
315
316 for (i = 0; i < bytes_in_fifo; i++)
317 i2c->rx_msg->buf[i2c->rx_pos++] =
318 xiic_getreg8(i2c, XIIC_DRR_REG_OFFSET);
319
320 xiic_setreg8(i2c, XIIC_RFD_REG_OFFSET,
321 (xiic_rx_space(i2c) > IIC_RX_FIFO_DEPTH) ?
322 IIC_RX_FIFO_DEPTH - 1 : xiic_rx_space(i2c) - 1);
323 }
324
xiic_tx_fifo_space(struct xiic_i2c * i2c)325 static int xiic_tx_fifo_space(struct xiic_i2c *i2c)
326 {
327 /* return the actual space left in the FIFO */
328 return IIC_TX_FIFO_DEPTH - xiic_getreg8(i2c, XIIC_TFO_REG_OFFSET) - 1;
329 }
330
xiic_fill_tx_fifo(struct xiic_i2c * i2c)331 static void xiic_fill_tx_fifo(struct xiic_i2c *i2c)
332 {
333 u8 fifo_space = xiic_tx_fifo_space(i2c);
334 int len = xiic_tx_space(i2c);
335
336 len = (len > fifo_space) ? fifo_space : len;
337
338 dev_dbg(i2c->adap.dev.parent, "%s entry, len: %d, fifo space: %d\n",
339 __func__, len, fifo_space);
340
341 while (len--) {
342 u16 data = i2c->tx_msg->buf[i2c->tx_pos++];
343 if ((xiic_tx_space(i2c) == 0) && (i2c->nmsgs == 1)) {
344 /* last message in transfer -> STOP */
345 data |= XIIC_TX_DYN_STOP_MASK;
346 dev_dbg(i2c->adap.dev.parent, "%s TX STOP\n", __func__);
347 }
348 xiic_setreg16(i2c, XIIC_DTR_REG_OFFSET, data);
349 }
350 }
351
xiic_wakeup(struct xiic_i2c * i2c,int code)352 static void xiic_wakeup(struct xiic_i2c *i2c, int code)
353 {
354 i2c->tx_msg = NULL;
355 i2c->rx_msg = NULL;
356 i2c->nmsgs = 0;
357 i2c->state = code;
358 wake_up(&i2c->wait);
359 }
360
xiic_process(int irq,void * dev_id)361 static irqreturn_t xiic_process(int irq, void *dev_id)
362 {
363 struct xiic_i2c *i2c = dev_id;
364 u32 pend, isr, ier;
365 u32 clr = 0;
366
367 /* Get the interrupt Status from the IPIF. There is no clearing of
368 * interrupts in the IPIF. Interrupts must be cleared at the source.
369 * To find which interrupts are pending; AND interrupts pending with
370 * interrupts masked.
371 */
372 spin_lock(&i2c->lock);
373 isr = xiic_getreg32(i2c, XIIC_IISR_OFFSET);
374 ier = xiic_getreg32(i2c, XIIC_IIER_OFFSET);
375 pend = isr & ier;
376
377 dev_dbg(i2c->adap.dev.parent, "%s: IER: 0x%x, ISR: 0x%x, pend: 0x%x\n",
378 __func__, ier, isr, pend);
379 dev_dbg(i2c->adap.dev.parent, "%s: SR: 0x%x, msg: %p, nmsgs: %d\n",
380 __func__, xiic_getreg8(i2c, XIIC_SR_REG_OFFSET),
381 i2c->tx_msg, i2c->nmsgs);
382
383
384 /* Service requesting interrupt */
385 if ((pend & XIIC_INTR_ARB_LOST_MASK) ||
386 ((pend & XIIC_INTR_TX_ERROR_MASK) &&
387 !(pend & XIIC_INTR_RX_FULL_MASK))) {
388 /* bus arbritration lost, or...
389 * Transmit error _OR_ RX completed
390 * if this happens when RX_FULL is not set
391 * this is probably a TX error
392 */
393
394 dev_dbg(i2c->adap.dev.parent, "%s error\n", __func__);
395
396 /* dynamic mode seem to suffer from problems if we just flushes
397 * fifos and the next message is a TX with len 0 (only addr)
398 * reset the IP instead of just flush fifos
399 */
400 xiic_reinit(i2c);
401
402 if (i2c->rx_msg)
403 xiic_wakeup(i2c, STATE_ERROR);
404 if (i2c->tx_msg)
405 xiic_wakeup(i2c, STATE_ERROR);
406 }
407 if (pend & XIIC_INTR_RX_FULL_MASK) {
408 /* Receive register/FIFO is full */
409
410 clr |= XIIC_INTR_RX_FULL_MASK;
411 if (!i2c->rx_msg) {
412 dev_dbg(i2c->adap.dev.parent,
413 "%s unexpexted RX IRQ\n", __func__);
414 xiic_clear_rx_fifo(i2c);
415 goto out;
416 }
417
418 xiic_read_rx(i2c);
419 if (xiic_rx_space(i2c) == 0) {
420 /* this is the last part of the message */
421 i2c->rx_msg = NULL;
422
423 /* also clear TX error if there (RX complete) */
424 clr |= (isr & XIIC_INTR_TX_ERROR_MASK);
425
426 dev_dbg(i2c->adap.dev.parent,
427 "%s end of message, nmsgs: %d\n",
428 __func__, i2c->nmsgs);
429
430 /* send next message if this wasn't the last,
431 * otherwise the transfer will be finialise when
432 * receiving the bus not busy interrupt
433 */
434 if (i2c->nmsgs > 1) {
435 i2c->nmsgs--;
436 i2c->tx_msg++;
437 dev_dbg(i2c->adap.dev.parent,
438 "%s will start next...\n", __func__);
439
440 __xiic_start_xfer(i2c);
441 }
442 }
443 }
444 if (pend & XIIC_INTR_BNB_MASK) {
445 /* IIC bus has transitioned to not busy */
446 clr |= XIIC_INTR_BNB_MASK;
447
448 /* The bus is not busy, disable BusNotBusy interrupt */
449 xiic_irq_dis(i2c, XIIC_INTR_BNB_MASK);
450
451 if (!i2c->tx_msg)
452 goto out;
453
454 if ((i2c->nmsgs == 1) && !i2c->rx_msg &&
455 xiic_tx_space(i2c) == 0)
456 xiic_wakeup(i2c, STATE_DONE);
457 else
458 xiic_wakeup(i2c, STATE_ERROR);
459 }
460 if (pend & (XIIC_INTR_TX_EMPTY_MASK | XIIC_INTR_TX_HALF_MASK)) {
461 /* Transmit register/FIFO is empty or ½ empty */
462
463 clr |= (pend &
464 (XIIC_INTR_TX_EMPTY_MASK | XIIC_INTR_TX_HALF_MASK));
465
466 if (!i2c->tx_msg) {
467 dev_dbg(i2c->adap.dev.parent,
468 "%s unexpexted TX IRQ\n", __func__);
469 goto out;
470 }
471
472 xiic_fill_tx_fifo(i2c);
473
474 /* current message sent and there is space in the fifo */
475 if (!xiic_tx_space(i2c) && xiic_tx_fifo_space(i2c) >= 2) {
476 dev_dbg(i2c->adap.dev.parent,
477 "%s end of message sent, nmsgs: %d\n",
478 __func__, i2c->nmsgs);
479 if (i2c->nmsgs > 1) {
480 i2c->nmsgs--;
481 i2c->tx_msg++;
482 __xiic_start_xfer(i2c);
483 } else {
484 xiic_irq_dis(i2c, XIIC_INTR_TX_HALF_MASK);
485
486 dev_dbg(i2c->adap.dev.parent,
487 "%s Got TX IRQ but no more to do...\n",
488 __func__);
489 }
490 } else if (!xiic_tx_space(i2c) && (i2c->nmsgs == 1))
491 /* current frame is sent and is last,
492 * make sure to disable tx half
493 */
494 xiic_irq_dis(i2c, XIIC_INTR_TX_HALF_MASK);
495 }
496 out:
497 dev_dbg(i2c->adap.dev.parent, "%s clr: 0x%x\n", __func__, clr);
498
499 xiic_setreg32(i2c, XIIC_IISR_OFFSET, clr);
500 spin_unlock(&i2c->lock);
501 return IRQ_HANDLED;
502 }
503
xiic_bus_busy(struct xiic_i2c * i2c)504 static int xiic_bus_busy(struct xiic_i2c *i2c)
505 {
506 u8 sr = xiic_getreg8(i2c, XIIC_SR_REG_OFFSET);
507
508 return (sr & XIIC_SR_BUS_BUSY_MASK) ? -EBUSY : 0;
509 }
510
xiic_busy(struct xiic_i2c * i2c)511 static int xiic_busy(struct xiic_i2c *i2c)
512 {
513 int tries = 3;
514 int err;
515
516 if (i2c->tx_msg)
517 return -EBUSY;
518
519 /* for instance if previous transfer was terminated due to TX error
520 * it might be that the bus is on it's way to become available
521 * give it at most 3 ms to wake
522 */
523 err = xiic_bus_busy(i2c);
524 while (err && tries--) {
525 msleep(1);
526 err = xiic_bus_busy(i2c);
527 }
528
529 return err;
530 }
531
xiic_start_recv(struct xiic_i2c * i2c)532 static void xiic_start_recv(struct xiic_i2c *i2c)
533 {
534 u8 rx_watermark;
535 struct i2c_msg *msg = i2c->rx_msg = i2c->tx_msg;
536 unsigned long flags;
537
538 /* Clear and enable Rx full interrupt. */
539 xiic_irq_clr_en(i2c, XIIC_INTR_RX_FULL_MASK | XIIC_INTR_TX_ERROR_MASK);
540
541 /* we want to get all but last byte, because the TX_ERROR IRQ is used
542 * to inidicate error ACK on the address, and negative ack on the last
543 * received byte, so to not mix them receive all but last.
544 * In the case where there is only one byte to receive
545 * we can check if ERROR and RX full is set at the same time
546 */
547 rx_watermark = msg->len;
548 if (rx_watermark > IIC_RX_FIFO_DEPTH)
549 rx_watermark = IIC_RX_FIFO_DEPTH;
550 xiic_setreg8(i2c, XIIC_RFD_REG_OFFSET, rx_watermark - 1);
551
552 local_irq_save(flags);
553 if (!(msg->flags & I2C_M_NOSTART))
554 /* write the address */
555 xiic_setreg16(i2c, XIIC_DTR_REG_OFFSET,
556 (msg->addr << 1) | XIIC_READ_OPERATION |
557 XIIC_TX_DYN_START_MASK);
558
559 xiic_irq_clr_en(i2c, XIIC_INTR_BNB_MASK);
560
561 xiic_setreg16(i2c, XIIC_DTR_REG_OFFSET,
562 msg->len | ((i2c->nmsgs == 1) ? XIIC_TX_DYN_STOP_MASK : 0));
563 local_irq_restore(flags);
564
565 if (i2c->nmsgs == 1)
566 /* very last, enable bus not busy as well */
567 xiic_irq_clr_en(i2c, XIIC_INTR_BNB_MASK);
568
569 /* the message is tx:ed */
570 i2c->tx_pos = msg->len;
571 }
572
xiic_start_send(struct xiic_i2c * i2c)573 static void xiic_start_send(struct xiic_i2c *i2c)
574 {
575 struct i2c_msg *msg = i2c->tx_msg;
576
577 xiic_irq_clr(i2c, XIIC_INTR_TX_ERROR_MASK);
578
579 dev_dbg(i2c->adap.dev.parent, "%s entry, msg: %p, len: %d",
580 __func__, msg, msg->len);
581 dev_dbg(i2c->adap.dev.parent, "%s entry, ISR: 0x%x, CR: 0x%x\n",
582 __func__, xiic_getreg32(i2c, XIIC_IISR_OFFSET),
583 xiic_getreg8(i2c, XIIC_CR_REG_OFFSET));
584
585 if (!(msg->flags & I2C_M_NOSTART)) {
586 /* write the address */
587 u16 data = ((msg->addr << 1) & 0xfe) | XIIC_WRITE_OPERATION |
588 XIIC_TX_DYN_START_MASK;
589 if ((i2c->nmsgs == 1) && msg->len == 0)
590 /* no data and last message -> add STOP */
591 data |= XIIC_TX_DYN_STOP_MASK;
592
593 xiic_setreg16(i2c, XIIC_DTR_REG_OFFSET, data);
594 }
595
596 xiic_fill_tx_fifo(i2c);
597
598 /* Clear any pending Tx empty, Tx Error and then enable them. */
599 xiic_irq_clr_en(i2c, XIIC_INTR_TX_EMPTY_MASK | XIIC_INTR_TX_ERROR_MASK |
600 XIIC_INTR_BNB_MASK);
601 }
602
xiic_isr(int irq,void * dev_id)603 static irqreturn_t xiic_isr(int irq, void *dev_id)
604 {
605 struct xiic_i2c *i2c = dev_id;
606 u32 pend, isr, ier;
607 irqreturn_t ret = IRQ_NONE;
608 /* Do not processes a devices interrupts if the device has no
609 * interrupts pending
610 */
611
612 dev_dbg(i2c->adap.dev.parent, "%s entry\n", __func__);
613
614 isr = xiic_getreg32(i2c, XIIC_IISR_OFFSET);
615 ier = xiic_getreg32(i2c, XIIC_IIER_OFFSET);
616 pend = isr & ier;
617 if (pend)
618 ret = IRQ_WAKE_THREAD;
619
620 return ret;
621 }
622
__xiic_start_xfer(struct xiic_i2c * i2c)623 static void __xiic_start_xfer(struct xiic_i2c *i2c)
624 {
625 int first = 1;
626 int fifo_space = xiic_tx_fifo_space(i2c);
627 dev_dbg(i2c->adap.dev.parent, "%s entry, msg: %p, fifos space: %d\n",
628 __func__, i2c->tx_msg, fifo_space);
629
630 if (!i2c->tx_msg)
631 return;
632
633 i2c->rx_pos = 0;
634 i2c->tx_pos = 0;
635 i2c->state = STATE_START;
636 while ((fifo_space >= 2) && (first || (i2c->nmsgs > 1))) {
637 if (!first) {
638 i2c->nmsgs--;
639 i2c->tx_msg++;
640 i2c->tx_pos = 0;
641 } else
642 first = 0;
643
644 if (i2c->tx_msg->flags & I2C_M_RD) {
645 /* we dont date putting several reads in the FIFO */
646 xiic_start_recv(i2c);
647 return;
648 } else {
649 xiic_start_send(i2c);
650 if (xiic_tx_space(i2c) != 0) {
651 /* the message could not be completely sent */
652 break;
653 }
654 }
655
656 fifo_space = xiic_tx_fifo_space(i2c);
657 }
658
659 /* there are more messages or the current one could not be completely
660 * put into the FIFO, also enable the half empty interrupt
661 */
662 if (i2c->nmsgs > 1 || xiic_tx_space(i2c))
663 xiic_irq_clr_en(i2c, XIIC_INTR_TX_HALF_MASK);
664
665 }
666
xiic_start_xfer(struct xiic_i2c * i2c)667 static void xiic_start_xfer(struct xiic_i2c *i2c)
668 {
669 spin_lock(&i2c->lock);
670 xiic_reinit(i2c);
671 __xiic_start_xfer(i2c);
672 spin_unlock(&i2c->lock);
673 }
674
xiic_xfer(struct i2c_adapter * adap,struct i2c_msg * msgs,int num)675 static int xiic_xfer(struct i2c_adapter *adap, struct i2c_msg *msgs, int num)
676 {
677 struct xiic_i2c *i2c = i2c_get_adapdata(adap);
678 int err;
679
680 dev_dbg(adap->dev.parent, "%s entry SR: 0x%x\n", __func__,
681 xiic_getreg8(i2c, XIIC_SR_REG_OFFSET));
682
683 err = xiic_busy(i2c);
684 if (err)
685 return err;
686
687 i2c->tx_msg = msgs;
688 i2c->nmsgs = num;
689
690 xiic_start_xfer(i2c);
691
692 if (wait_event_timeout(i2c->wait, (i2c->state == STATE_ERROR) ||
693 (i2c->state == STATE_DONE), HZ))
694 return (i2c->state == STATE_DONE) ? num : -EIO;
695 else {
696 i2c->tx_msg = NULL;
697 i2c->rx_msg = NULL;
698 i2c->nmsgs = 0;
699 return -ETIMEDOUT;
700 }
701 }
702
xiic_func(struct i2c_adapter * adap)703 static u32 xiic_func(struct i2c_adapter *adap)
704 {
705 return I2C_FUNC_I2C | I2C_FUNC_SMBUS_EMUL;
706 }
707
708 static const struct i2c_algorithm xiic_algorithm = {
709 .master_xfer = xiic_xfer,
710 .functionality = xiic_func,
711 };
712
713 static struct i2c_adapter xiic_adapter = {
714 .owner = THIS_MODULE,
715 .name = DRIVER_NAME,
716 .class = I2C_CLASS_DEPRECATED,
717 .algo = &xiic_algorithm,
718 };
719
720
xiic_i2c_probe(struct platform_device * pdev)721 static int xiic_i2c_probe(struct platform_device *pdev)
722 {
723 struct xiic_i2c *i2c;
724 struct xiic_i2c_platform_data *pdata;
725 struct resource *res;
726 int ret, irq;
727 u8 i;
728 u32 sr;
729
730 i2c = devm_kzalloc(&pdev->dev, sizeof(*i2c), GFP_KERNEL);
731 if (!i2c)
732 return -ENOMEM;
733
734 res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
735 i2c->base = devm_ioremap_resource(&pdev->dev, res);
736 if (IS_ERR(i2c->base))
737 return PTR_ERR(i2c->base);
738
739 irq = platform_get_irq(pdev, 0);
740 if (irq < 0)
741 return irq;
742
743 pdata = dev_get_platdata(&pdev->dev);
744
745 /* hook up driver to tree */
746 platform_set_drvdata(pdev, i2c);
747 i2c->adap = xiic_adapter;
748 i2c_set_adapdata(&i2c->adap, i2c);
749 i2c->adap.dev.parent = &pdev->dev;
750 i2c->adap.dev.of_node = pdev->dev.of_node;
751
752 spin_lock_init(&i2c->lock);
753 init_waitqueue_head(&i2c->wait);
754
755 ret = devm_request_threaded_irq(&pdev->dev, irq, xiic_isr,
756 xiic_process, IRQF_ONESHOT,
757 pdev->name, i2c);
758
759 if (ret < 0) {
760 dev_err(&pdev->dev, "Cannot claim IRQ\n");
761 return ret;
762 }
763
764 /*
765 * Detect endianness
766 * Try to reset the TX FIFO. Then check the EMPTY flag. If it is not
767 * set, assume that the endianness was wrong and swap.
768 */
769 i2c->endianness = LITTLE;
770 xiic_setreg32(i2c, XIIC_CR_REG_OFFSET, XIIC_CR_TX_FIFO_RESET_MASK);
771 /* Reset is cleared in xiic_reinit */
772 sr = xiic_getreg32(i2c, XIIC_SR_REG_OFFSET);
773 if (!(sr & XIIC_SR_TX_FIFO_EMPTY_MASK))
774 i2c->endianness = BIG;
775
776 xiic_reinit(i2c);
777
778 /* add i2c adapter to i2c tree */
779 ret = i2c_add_adapter(&i2c->adap);
780 if (ret) {
781 dev_err(&pdev->dev, "Failed to add adapter\n");
782 xiic_deinit(i2c);
783 return ret;
784 }
785
786 if (pdata) {
787 /* add in known devices to the bus */
788 for (i = 0; i < pdata->num_devices; i++)
789 i2c_new_device(&i2c->adap, pdata->devices + i);
790 }
791
792 return 0;
793 }
794
xiic_i2c_remove(struct platform_device * pdev)795 static int xiic_i2c_remove(struct platform_device *pdev)
796 {
797 struct xiic_i2c *i2c = platform_get_drvdata(pdev);
798
799 /* remove adapter & data */
800 i2c_del_adapter(&i2c->adap);
801
802 xiic_deinit(i2c);
803
804 return 0;
805 }
806
807 #if defined(CONFIG_OF)
808 static const struct of_device_id xiic_of_match[] = {
809 { .compatible = "xlnx,xps-iic-2.00.a", },
810 {},
811 };
812 MODULE_DEVICE_TABLE(of, xiic_of_match);
813 #endif
814
815 static struct platform_driver xiic_i2c_driver = {
816 .probe = xiic_i2c_probe,
817 .remove = xiic_i2c_remove,
818 .driver = {
819 .name = DRIVER_NAME,
820 .of_match_table = of_match_ptr(xiic_of_match),
821 },
822 };
823
824 module_platform_driver(xiic_i2c_driver);
825
826 MODULE_AUTHOR("info@mocean-labs.com");
827 MODULE_DESCRIPTION("Xilinx I2C bus driver");
828 MODULE_LICENSE("GPL v2");
829 MODULE_ALIAS("platform:"DRIVER_NAME);
830