1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * Synopsys DesignWare I2C adapter driver (master only).
5 * Based on the TI DAVINCI I2C adapter driver.
7 * Copyright (C) 2006 Texas Instruments.
8 * Copyright (C) 2007 MontaVista Software Inc.
9 * Copyright (C) 2009 Provigent Ltd.
11 #include <linux/delay.h>
12 #include <linux/err.h>
13 #include <linux/errno.h>
14 #include <linux/export.h>
15 #include <linux/gpio/consumer.h>
16 #include <linux/i2c.h>
17 #include <linux/interrupt.h>
19 #include <linux/module.h>
20 #include <linux/pm_runtime.h>
21 #include <linux/regmap.h>
22 #include <linux/reset.h>
24 #include "i2c-designware-core.h"
26 #define AMD_TIMEOUT_MIN_US 25
27 #define AMD_TIMEOUT_MAX_US 250
28 #define AMD_MASTERCFG_MASK GENMASK(15, 0)
30 static void i2c_dw_configure_fifo_master(struct dw_i2c_dev *dev)
32 /* Configure Tx/Rx FIFO threshold levels */
33 regmap_write(dev->map, DW_IC_TX_TL, dev->tx_fifo_depth / 2);
34 regmap_write(dev->map, DW_IC_RX_TL, 0);
36 /* Configure the I2C master */
37 regmap_write(dev->map, DW_IC_CON, dev->master_cfg);
40 static int i2c_dw_set_timings_master(struct dw_i2c_dev *dev)
43 u32 sda_falling_time, scl_falling_time;
44 struct i2c_timings *t = &dev->timings;
45 const char *fp_str = "";
49 ret = i2c_dw_acquire_lock(dev);
53 ret = regmap_read(dev->map, DW_IC_COMP_PARAM_1, &comp_param1);
54 i2c_dw_release_lock(dev);
58 /* Set standard and fast speed dividers for high/low periods */
59 sda_falling_time = t->sda_fall_ns ?: 300; /* ns */
60 scl_falling_time = t->scl_fall_ns ?: 300; /* ns */
62 /* Calculate SCL timing parameters for standard mode if not set */
63 if (!dev->ss_hcnt || !dev->ss_lcnt) {
64 ic_clk = i2c_dw_clk_rate(dev);
66 i2c_dw_scl_hcnt(ic_clk,
67 4000, /* tHD;STA = tHIGH = 4.0 us */
69 0, /* 0: DW default, 1: Ideal */
72 i2c_dw_scl_lcnt(ic_clk,
73 4700, /* tLOW = 4.7 us */
77 dev_dbg(dev->dev, "Standard Mode HCNT:LCNT = %d:%d\n",
78 dev->ss_hcnt, dev->ss_lcnt);
81 * Set SCL timing parameters for fast mode or fast mode plus. Only
82 * difference is the timing parameter values since the registers are
85 if (t->bus_freq_hz == I2C_MAX_FAST_MODE_PLUS_FREQ) {
87 * Check are Fast Mode Plus parameters available. Calculate
88 * SCL timing parameters for Fast Mode Plus if not set.
90 if (dev->fp_hcnt && dev->fp_lcnt) {
91 dev->fs_hcnt = dev->fp_hcnt;
92 dev->fs_lcnt = dev->fp_lcnt;
94 ic_clk = i2c_dw_clk_rate(dev);
96 i2c_dw_scl_hcnt(ic_clk,
97 260, /* tHIGH = 260 ns */
102 i2c_dw_scl_lcnt(ic_clk,
103 500, /* tLOW = 500 ns */
110 * Calculate SCL timing parameters for fast mode if not set. They are
111 * needed also in high speed mode.
113 if (!dev->fs_hcnt || !dev->fs_lcnt) {
114 ic_clk = i2c_dw_clk_rate(dev);
116 i2c_dw_scl_hcnt(ic_clk,
117 600, /* tHD;STA = tHIGH = 0.6 us */
119 0, /* 0: DW default, 1: Ideal */
122 i2c_dw_scl_lcnt(ic_clk,
123 1300, /* tLOW = 1.3 us */
127 dev_dbg(dev->dev, "Fast Mode%s HCNT:LCNT = %d:%d\n",
128 fp_str, dev->fs_hcnt, dev->fs_lcnt);
130 /* Check is high speed possible and fall back to fast mode if not */
131 if ((dev->master_cfg & DW_IC_CON_SPEED_MASK) ==
132 DW_IC_CON_SPEED_HIGH) {
133 if ((comp_param1 & DW_IC_COMP_PARAM_1_SPEED_MODE_MASK)
134 != DW_IC_COMP_PARAM_1_SPEED_MODE_HIGH) {
135 dev_err(dev->dev, "High Speed not supported!\n");
136 t->bus_freq_hz = I2C_MAX_FAST_MODE_FREQ;
137 dev->master_cfg &= ~DW_IC_CON_SPEED_MASK;
138 dev->master_cfg |= DW_IC_CON_SPEED_FAST;
141 } else if (!dev->hs_hcnt || !dev->hs_lcnt) {
142 ic_clk = i2c_dw_clk_rate(dev);
144 i2c_dw_scl_hcnt(ic_clk,
145 160, /* tHIGH = 160 ns */
150 i2c_dw_scl_lcnt(ic_clk,
151 320, /* tLOW = 320 ns */
155 dev_dbg(dev->dev, "High Speed Mode HCNT:LCNT = %d:%d\n",
156 dev->hs_hcnt, dev->hs_lcnt);
159 ret = i2c_dw_set_sda_hold(dev);
163 dev_dbg(dev->dev, "Bus speed: %s\n", i2c_freq_mode_string(t->bus_freq_hz));
168 * i2c_dw_init_master() - Initialize the designware I2C master hardware
169 * @dev: device private data
171 * This functions configures and enables the I2C master.
172 * This function is called during I2C init function, and in case of timeout at
175 static int i2c_dw_init_master(struct dw_i2c_dev *dev)
179 ret = i2c_dw_acquire_lock(dev);
183 /* Disable the adapter */
184 __i2c_dw_disable(dev);
186 /* Write standard speed timing parameters */
187 regmap_write(dev->map, DW_IC_SS_SCL_HCNT, dev->ss_hcnt);
188 regmap_write(dev->map, DW_IC_SS_SCL_LCNT, dev->ss_lcnt);
190 /* Write fast mode/fast mode plus timing parameters */
191 regmap_write(dev->map, DW_IC_FS_SCL_HCNT, dev->fs_hcnt);
192 regmap_write(dev->map, DW_IC_FS_SCL_LCNT, dev->fs_lcnt);
194 /* Write high speed timing parameters if supported */
195 if (dev->hs_hcnt && dev->hs_lcnt) {
196 regmap_write(dev->map, DW_IC_HS_SCL_HCNT, dev->hs_hcnt);
197 regmap_write(dev->map, DW_IC_HS_SCL_LCNT, dev->hs_lcnt);
200 /* Write SDA hold time if supported */
201 if (dev->sda_hold_time)
202 regmap_write(dev->map, DW_IC_SDA_HOLD, dev->sda_hold_time);
204 i2c_dw_configure_fifo_master(dev);
205 i2c_dw_release_lock(dev);
210 static void i2c_dw_xfer_init(struct dw_i2c_dev *dev)
212 struct i2c_msg *msgs = dev->msgs;
213 u32 ic_con = 0, ic_tar = 0;
216 /* Disable the adapter */
217 __i2c_dw_disable(dev);
219 /* If the slave address is ten bit address, enable 10BITADDR */
220 if (msgs[dev->msg_write_idx].flags & I2C_M_TEN) {
221 ic_con = DW_IC_CON_10BITADDR_MASTER;
223 * If I2C_DYNAMIC_TAR_UPDATE is set, the 10-bit addressing
224 * mode has to be enabled via bit 12 of IC_TAR register.
225 * We set it always as I2C_DYNAMIC_TAR_UPDATE can't be
226 * detected from registers.
228 ic_tar = DW_IC_TAR_10BITADDR_MASTER;
231 regmap_update_bits(dev->map, DW_IC_CON, DW_IC_CON_10BITADDR_MASTER,
235 * Set the slave (target) address and enable 10-bit addressing mode
238 regmap_write(dev->map, DW_IC_TAR,
239 msgs[dev->msg_write_idx].addr | ic_tar);
241 /* Enforce disabled interrupts (due to HW issues) */
242 i2c_dw_disable_int(dev);
244 /* Enable the adapter */
245 __i2c_dw_enable(dev);
247 /* Dummy read to avoid the register getting stuck on Bay Trail */
248 regmap_read(dev->map, DW_IC_ENABLE_STATUS, &dummy);
250 /* Clear and enable interrupts */
251 regmap_read(dev->map, DW_IC_CLR_INTR, &dummy);
252 regmap_write(dev->map, DW_IC_INTR_MASK, DW_IC_INTR_MASTER_MASK);
255 static int i2c_dw_check_stopbit(struct dw_i2c_dev *dev)
260 ret = regmap_read_poll_timeout(dev->map, DW_IC_INTR_STAT, val,
261 !(val & DW_IC_INTR_STOP_DET),
264 dev_err(dev->dev, "i2c timeout error %d\n", ret);
269 static int i2c_dw_status(struct dw_i2c_dev *dev)
273 status = i2c_dw_wait_bus_not_busy(dev);
277 return i2c_dw_check_stopbit(dev);
281 * Initiate and continue master read/write transaction with polling
282 * based transfer routine afterward write messages into the Tx buffer.
284 static int amd_i2c_dw_xfer_quirk(struct i2c_adapter *adap, struct i2c_msg *msgs, int num_msgs)
286 struct dw_i2c_dev *dev = i2c_get_adapdata(adap);
287 int msg_wrt_idx, msg_itr_lmt, buf_len, data_idx;
293 * In order to enable the interrupt for UCSI i.e. AMD NAVI GPU card,
294 * it is mandatory to set the right value in specific register
295 * (offset:0x474) as per the hardware IP specification.
297 regmap_write(dev->map, AMD_UCSI_INTR_REG, AMD_UCSI_INTR_EN);
300 dev->msgs_num = num_msgs;
301 i2c_dw_xfer_init(dev);
302 i2c_dw_disable_int(dev);
304 /* Initiate messages read/write transaction */
305 for (msg_wrt_idx = 0; msg_wrt_idx < num_msgs; msg_wrt_idx++) {
306 tx_buf = msgs[msg_wrt_idx].buf;
307 buf_len = msgs[msg_wrt_idx].len;
309 if (!(msgs[msg_wrt_idx].flags & I2C_M_RD))
310 regmap_write(dev->map, DW_IC_TX_TL, buf_len - 1);
312 * Initiate the i2c read/write transaction of buffer length,
313 * and poll for bus busy status. For the last message transfer,
314 * update the command with stopbit enable.
316 for (msg_itr_lmt = buf_len; msg_itr_lmt > 0; msg_itr_lmt--) {
317 if (msg_wrt_idx == num_msgs - 1 && msg_itr_lmt == 1)
320 if (msgs[msg_wrt_idx].flags & I2C_M_RD) {
321 /* Due to hardware bug, need to write the same command twice. */
322 regmap_write(dev->map, DW_IC_DATA_CMD, 0x100);
323 regmap_write(dev->map, DW_IC_DATA_CMD, 0x100 | cmd);
325 regmap_write(dev->map, DW_IC_TX_TL, 2 * (buf_len - 1));
326 regmap_write(dev->map, DW_IC_RX_TL, 2 * (buf_len - 1));
328 * Need to check the stop bit. However, it cannot be
329 * detected from the registers so we check it always
330 * when read/write the last byte.
332 status = i2c_dw_status(dev);
336 for (data_idx = 0; data_idx < buf_len; data_idx++) {
337 regmap_read(dev->map, DW_IC_DATA_CMD, &val);
338 tx_buf[data_idx] = val;
340 status = i2c_dw_check_stopbit(dev);
345 regmap_write(dev->map, DW_IC_DATA_CMD, *tx_buf++ | cmd);
346 usleep_range(AMD_TIMEOUT_MIN_US, AMD_TIMEOUT_MAX_US);
349 status = i2c_dw_check_stopbit(dev);
358 * Initiate (and continue) low level master read/write transaction.
359 * This function is only called from i2c_dw_isr, and pumping i2c_msg
360 * messages into the tx buffer. Even if the size of i2c_msg data is
361 * longer than the size of the tx buffer, it handles everything.
364 i2c_dw_xfer_msg(struct dw_i2c_dev *dev)
366 struct i2c_msg *msgs = dev->msgs;
368 int tx_limit, rx_limit;
369 u32 addr = msgs[dev->msg_write_idx].addr;
370 u32 buf_len = dev->tx_buf_len;
371 u8 *buf = dev->tx_buf;
372 bool need_restart = false;
375 intr_mask = DW_IC_INTR_MASTER_MASK;
377 for (; dev->msg_write_idx < dev->msgs_num; dev->msg_write_idx++) {
378 u32 flags = msgs[dev->msg_write_idx].flags;
381 * If target address has changed, we need to
382 * reprogram the target address in the I2C
383 * adapter when we are done with this transfer.
385 if (msgs[dev->msg_write_idx].addr != addr) {
387 "%s: invalid target address\n", __func__);
388 dev->msg_err = -EINVAL;
392 if (!(dev->status & STATUS_WRITE_IN_PROGRESS)) {
394 buf = msgs[dev->msg_write_idx].buf;
395 buf_len = msgs[dev->msg_write_idx].len;
397 /* If both IC_EMPTYFIFO_HOLD_MASTER_EN and
398 * IC_RESTART_EN are set, we must manually
399 * set restart bit between messages.
401 if ((dev->master_cfg & DW_IC_CON_RESTART_EN) &&
402 (dev->msg_write_idx > 0))
406 regmap_read(dev->map, DW_IC_TXFLR, &flr);
407 tx_limit = dev->tx_fifo_depth - flr;
409 regmap_read(dev->map, DW_IC_RXFLR, &flr);
410 rx_limit = dev->rx_fifo_depth - flr;
412 while (buf_len > 0 && tx_limit > 0 && rx_limit > 0) {
416 * If IC_EMPTYFIFO_HOLD_MASTER_EN is set we must
417 * manually set the stop bit. However, it cannot be
418 * detected from the registers so we set it always
419 * when writing/reading the last byte.
423 * i2c-core always sets the buffer length of
424 * I2C_FUNC_SMBUS_BLOCK_DATA to 1. The length will
425 * be adjusted when receiving the first byte.
426 * Thus we can't stop the transaction here.
428 if (dev->msg_write_idx == dev->msgs_num - 1 &&
429 buf_len == 1 && !(flags & I2C_M_RECV_LEN))
434 need_restart = false;
437 if (msgs[dev->msg_write_idx].flags & I2C_M_RD) {
439 /* Avoid rx buffer overrun */
440 if (dev->rx_outstanding >= dev->rx_fifo_depth)
443 regmap_write(dev->map, DW_IC_DATA_CMD,
446 dev->rx_outstanding++;
448 regmap_write(dev->map, DW_IC_DATA_CMD,
451 tx_limit--; buf_len--;
455 dev->tx_buf_len = buf_len;
458 * Because we don't know the buffer length in the
459 * I2C_FUNC_SMBUS_BLOCK_DATA case, we can't stop
460 * the transaction here.
462 if (buf_len > 0 || flags & I2C_M_RECV_LEN) {
463 /* more bytes to be written */
464 dev->status |= STATUS_WRITE_IN_PROGRESS;
467 dev->status &= ~STATUS_WRITE_IN_PROGRESS;
471 * If i2c_msg index search is completed, we don't need TX_EMPTY
472 * interrupt any more.
474 if (dev->msg_write_idx == dev->msgs_num)
475 intr_mask &= ~DW_IC_INTR_TX_EMPTY;
480 regmap_write(dev->map, DW_IC_INTR_MASK, intr_mask);
484 i2c_dw_recv_len(struct dw_i2c_dev *dev, u8 len)
486 struct i2c_msg *msgs = dev->msgs;
487 u32 flags = msgs[dev->msg_read_idx].flags;
490 * Adjust the buffer length and mask the flag
491 * after receiving the first byte.
493 len += (flags & I2C_CLIENT_PEC) ? 2 : 1;
494 dev->tx_buf_len = len - min_t(u8, len, dev->rx_outstanding);
495 msgs[dev->msg_read_idx].len = len;
496 msgs[dev->msg_read_idx].flags &= ~I2C_M_RECV_LEN;
502 i2c_dw_read(struct dw_i2c_dev *dev)
504 struct i2c_msg *msgs = dev->msgs;
505 unsigned int rx_valid;
507 for (; dev->msg_read_idx < dev->msgs_num; dev->msg_read_idx++) {
511 if (!(msgs[dev->msg_read_idx].flags & I2C_M_RD))
514 if (!(dev->status & STATUS_READ_IN_PROGRESS)) {
515 len = msgs[dev->msg_read_idx].len;
516 buf = msgs[dev->msg_read_idx].buf;
518 len = dev->rx_buf_len;
522 regmap_read(dev->map, DW_IC_RXFLR, &rx_valid);
524 for (; len > 0 && rx_valid > 0; len--, rx_valid--) {
525 u32 flags = msgs[dev->msg_read_idx].flags;
527 regmap_read(dev->map, DW_IC_DATA_CMD, &tmp);
528 /* Ensure length byte is a valid value */
529 if (flags & I2C_M_RECV_LEN &&
530 (tmp & DW_IC_DATA_CMD_DAT) <= I2C_SMBUS_BLOCK_MAX && tmp > 0) {
531 len = i2c_dw_recv_len(dev, tmp);
534 dev->rx_outstanding--;
538 dev->status |= STATUS_READ_IN_PROGRESS;
539 dev->rx_buf_len = len;
543 dev->status &= ~STATUS_READ_IN_PROGRESS;
548 * Prepare controller for a transaction and call i2c_dw_xfer_msg.
551 i2c_dw_xfer(struct i2c_adapter *adap, struct i2c_msg msgs[], int num)
553 struct dw_i2c_dev *dev = i2c_get_adapdata(adap);
556 dev_dbg(dev->dev, "%s: msgs: %d\n", __func__, num);
558 pm_runtime_get_sync(dev->dev);
561 * Initiate I2C message transfer when AMD NAVI GPU card is enabled,
562 * As it is polling based transfer mechanism, which does not support
563 * interrupt based functionalities of existing DesignWare driver.
565 if ((dev->flags & MODEL_MASK) == MODEL_AMD_NAVI_GPU) {
566 ret = amd_i2c_dw_xfer_quirk(adap, msgs, num);
570 reinit_completion(&dev->cmd_complete);
574 dev->msg_write_idx = 0;
575 dev->msg_read_idx = 0;
577 dev->status = STATUS_IDLE;
578 dev->abort_source = 0;
579 dev->rx_outstanding = 0;
581 ret = i2c_dw_acquire_lock(dev);
585 ret = i2c_dw_wait_bus_not_busy(dev);
589 /* Start the transfers */
590 i2c_dw_xfer_init(dev);
592 /* Wait for tx to complete */
593 if (!wait_for_completion_timeout(&dev->cmd_complete, adap->timeout)) {
594 dev_err(dev->dev, "controller timed out\n");
595 /* i2c_dw_init implicitly disables the adapter */
596 i2c_recover_bus(&dev->adapter);
597 i2c_dw_init_master(dev);
603 * We must disable the adapter before returning and signaling the end
604 * of the current transfer. Otherwise the hardware might continue
605 * generating interrupts which in turn causes a race condition with
606 * the following transfer. Needs some more investigation if the
607 * additional interrupts are a hardware bug or this driver doesn't
608 * handle them correctly yet.
610 __i2c_dw_disable_nowait(dev);
618 if (likely(!dev->cmd_err && !dev->status)) {
623 /* We have an error */
624 if (dev->cmd_err == DW_IC_ERR_TX_ABRT) {
625 ret = i2c_dw_handle_tx_abort(dev);
631 "transfer terminated early - interrupt latency too high?\n");
636 i2c_dw_release_lock(dev);
639 pm_runtime_mark_last_busy(dev->dev);
640 pm_runtime_put_autosuspend(dev->dev);
645 static const struct i2c_algorithm i2c_dw_algo = {
646 .master_xfer = i2c_dw_xfer,
647 .functionality = i2c_dw_func,
650 static const struct i2c_adapter_quirks i2c_dw_quirks = {
651 .flags = I2C_AQ_NO_ZERO_LEN,
654 static u32 i2c_dw_read_clear_intrbits(struct dw_i2c_dev *dev)
659 * The IC_INTR_STAT register just indicates "enabled" interrupts.
660 * The unmasked raw version of interrupt status bits is available
661 * in the IC_RAW_INTR_STAT register.
664 * stat = readl(IC_INTR_STAT);
666 * stat = readl(IC_RAW_INTR_STAT) & readl(IC_INTR_MASK);
668 * The raw version might be useful for debugging purposes.
670 regmap_read(dev->map, DW_IC_INTR_STAT, &stat);
673 * Do not use the IC_CLR_INTR register to clear interrupts, or
674 * you'll miss some interrupts, triggered during the period from
675 * readl(IC_INTR_STAT) to readl(IC_CLR_INTR).
677 * Instead, use the separately-prepared IC_CLR_* registers.
679 if (stat & DW_IC_INTR_RX_UNDER)
680 regmap_read(dev->map, DW_IC_CLR_RX_UNDER, &dummy);
681 if (stat & DW_IC_INTR_RX_OVER)
682 regmap_read(dev->map, DW_IC_CLR_RX_OVER, &dummy);
683 if (stat & DW_IC_INTR_TX_OVER)
684 regmap_read(dev->map, DW_IC_CLR_TX_OVER, &dummy);
685 if (stat & DW_IC_INTR_RD_REQ)
686 regmap_read(dev->map, DW_IC_CLR_RD_REQ, &dummy);
687 if (stat & DW_IC_INTR_TX_ABRT) {
689 * The IC_TX_ABRT_SOURCE register is cleared whenever
690 * the IC_CLR_TX_ABRT is read. Preserve it beforehand.
692 regmap_read(dev->map, DW_IC_TX_ABRT_SOURCE, &dev->abort_source);
693 regmap_read(dev->map, DW_IC_CLR_TX_ABRT, &dummy);
695 if (stat & DW_IC_INTR_RX_DONE)
696 regmap_read(dev->map, DW_IC_CLR_RX_DONE, &dummy);
697 if (stat & DW_IC_INTR_ACTIVITY)
698 regmap_read(dev->map, DW_IC_CLR_ACTIVITY, &dummy);
699 if ((stat & DW_IC_INTR_STOP_DET) &&
700 ((dev->rx_outstanding == 0) || (stat & DW_IC_INTR_RX_FULL)))
701 regmap_read(dev->map, DW_IC_CLR_STOP_DET, &dummy);
702 if (stat & DW_IC_INTR_START_DET)
703 regmap_read(dev->map, DW_IC_CLR_START_DET, &dummy);
704 if (stat & DW_IC_INTR_GEN_CALL)
705 regmap_read(dev->map, DW_IC_CLR_GEN_CALL, &dummy);
711 * Interrupt service routine. This gets called whenever an I2C master interrupt
714 static int i2c_dw_irq_handler_master(struct dw_i2c_dev *dev)
718 stat = i2c_dw_read_clear_intrbits(dev);
719 if (stat & DW_IC_INTR_TX_ABRT) {
720 dev->cmd_err |= DW_IC_ERR_TX_ABRT;
721 dev->status = STATUS_IDLE;
722 dev->rx_outstanding = 0;
725 * Anytime TX_ABRT is set, the contents of the tx/rx
726 * buffers are flushed. Make sure to skip them.
728 regmap_write(dev->map, DW_IC_INTR_MASK, 0);
732 if (stat & DW_IC_INTR_RX_FULL)
735 if (stat & DW_IC_INTR_TX_EMPTY)
736 i2c_dw_xfer_msg(dev);
739 * No need to modify or disable the interrupt mask here.
740 * i2c_dw_xfer_msg() will take care of it according to
741 * the current transmit status.
745 if (((stat & (DW_IC_INTR_TX_ABRT | DW_IC_INTR_STOP_DET)) || dev->msg_err) &&
746 (dev->rx_outstanding == 0))
747 complete(&dev->cmd_complete);
748 else if (unlikely(dev->flags & ACCESS_INTR_MASK)) {
749 /* Workaround to trigger pending interrupt */
750 regmap_read(dev->map, DW_IC_INTR_MASK, &stat);
751 i2c_dw_disable_int(dev);
752 regmap_write(dev->map, DW_IC_INTR_MASK, stat);
758 static irqreturn_t i2c_dw_isr(int this_irq, void *dev_id)
760 struct dw_i2c_dev *dev = dev_id;
763 regmap_read(dev->map, DW_IC_ENABLE, &enabled);
764 regmap_read(dev->map, DW_IC_RAW_INTR_STAT, &stat);
765 dev_dbg(dev->dev, "enabled=%#x stat=%#x\n", enabled, stat);
766 if (!enabled || !(stat & ~DW_IC_INTR_ACTIVITY))
769 i2c_dw_irq_handler_master(dev);
774 void i2c_dw_configure_master(struct dw_i2c_dev *dev)
776 struct i2c_timings *t = &dev->timings;
778 dev->functionality = I2C_FUNC_10BIT_ADDR | DW_IC_DEFAULT_FUNCTIONALITY;
780 dev->master_cfg = DW_IC_CON_MASTER | DW_IC_CON_SLAVE_DISABLE |
781 DW_IC_CON_RESTART_EN;
783 dev->mode = DW_IC_MASTER;
785 switch (t->bus_freq_hz) {
786 case I2C_MAX_STANDARD_MODE_FREQ:
787 dev->master_cfg |= DW_IC_CON_SPEED_STD;
789 case I2C_MAX_HIGH_SPEED_MODE_FREQ:
790 dev->master_cfg |= DW_IC_CON_SPEED_HIGH;
793 dev->master_cfg |= DW_IC_CON_SPEED_FAST;
796 EXPORT_SYMBOL_GPL(i2c_dw_configure_master);
798 static void i2c_dw_prepare_recovery(struct i2c_adapter *adap)
800 struct dw_i2c_dev *dev = i2c_get_adapdata(adap);
803 reset_control_assert(dev->rst);
804 i2c_dw_prepare_clk(dev, false);
807 static void i2c_dw_unprepare_recovery(struct i2c_adapter *adap)
809 struct dw_i2c_dev *dev = i2c_get_adapdata(adap);
811 i2c_dw_prepare_clk(dev, true);
812 reset_control_deassert(dev->rst);
813 i2c_dw_init_master(dev);
816 static int i2c_dw_init_recovery_info(struct dw_i2c_dev *dev)
818 struct i2c_bus_recovery_info *rinfo = &dev->rinfo;
819 struct i2c_adapter *adap = &dev->adapter;
820 struct gpio_desc *gpio;
822 gpio = devm_gpiod_get_optional(dev->dev, "scl", GPIOD_OUT_HIGH);
823 if (IS_ERR_OR_NULL(gpio))
824 return PTR_ERR_OR_ZERO(gpio);
826 rinfo->scl_gpiod = gpio;
828 gpio = devm_gpiod_get_optional(dev->dev, "sda", GPIOD_IN);
830 return PTR_ERR(gpio);
831 rinfo->sda_gpiod = gpio;
833 rinfo->recover_bus = i2c_generic_scl_recovery;
834 rinfo->prepare_recovery = i2c_dw_prepare_recovery;
835 rinfo->unprepare_recovery = i2c_dw_unprepare_recovery;
836 adap->bus_recovery_info = rinfo;
838 dev_info(dev->dev, "running with gpio recovery mode! scl%s",
839 rinfo->sda_gpiod ? ",sda" : "");
844 static int amd_i2c_adap_quirk(struct dw_i2c_dev *dev)
846 struct i2c_adapter *adap = &dev->adapter;
849 pm_runtime_get_noresume(dev->dev);
850 ret = i2c_add_numbered_adapter(adap);
852 dev_err(dev->dev, "Failed to add adapter: %d\n", ret);
853 pm_runtime_put_noidle(dev->dev);
858 int i2c_dw_probe_master(struct dw_i2c_dev *dev)
860 struct i2c_adapter *adap = &dev->adapter;
861 unsigned long irq_flags;
864 init_completion(&dev->cmd_complete);
866 dev->init = i2c_dw_init_master;
867 dev->disable = i2c_dw_disable;
868 dev->disable_int = i2c_dw_disable_int;
870 ret = i2c_dw_init_regmap(dev);
874 ret = i2c_dw_set_timings_master(dev);
878 ret = i2c_dw_set_fifo_size(dev);
882 ret = dev->init(dev);
886 snprintf(adap->name, sizeof(adap->name),
887 "Synopsys DesignWare I2C adapter");
889 adap->algo = &i2c_dw_algo;
890 adap->quirks = &i2c_dw_quirks;
891 adap->dev.parent = dev->dev;
892 i2c_set_adapdata(adap, dev);
894 if ((dev->flags & MODEL_MASK) == MODEL_AMD_NAVI_GPU)
895 return amd_i2c_adap_quirk(dev);
897 if (dev->flags & ACCESS_NO_IRQ_SUSPEND) {
898 irq_flags = IRQF_NO_SUSPEND;
900 irq_flags = IRQF_SHARED | IRQF_COND_SUSPEND;
903 ret = i2c_dw_acquire_lock(dev);
907 i2c_dw_disable_int(dev);
908 i2c_dw_release_lock(dev);
910 ret = devm_request_irq(dev->dev, dev->irq, i2c_dw_isr, irq_flags,
911 dev_name(dev->dev), dev);
913 dev_err(dev->dev, "failure requesting irq %i: %d\n",
918 ret = i2c_dw_init_recovery_info(dev);
923 * Increment PM usage count during adapter registration in order to
924 * avoid possible spurious runtime suspend when adapter device is
925 * registered to the device core and immediate resume in case bus has
926 * registered I2C slaves that do I2C transfers in their probe.
928 pm_runtime_get_noresume(dev->dev);
929 ret = i2c_add_numbered_adapter(adap);
931 dev_err(dev->dev, "failure adding adapter: %d\n", ret);
932 pm_runtime_put_noidle(dev->dev);
936 EXPORT_SYMBOL_GPL(i2c_dw_probe_master);
938 MODULE_DESCRIPTION("Synopsys DesignWare I2C bus master adapter");
939 MODULE_LICENSE("GPL");