1 /* SPDX-License-Identifier: GPL-2.0+ */
2 /* Copyright (C) 2018 Microchip Technology Inc. */
4 #include <linux/module.h>
6 #include <linux/netdevice.h>
7 #include <linux/etherdevice.h>
8 #include <linux/crc32.h>
9 #include <linux/microchipphy.h>
10 #include <linux/net_tstamp.h>
11 #include <linux/phy.h>
12 #include <linux/rtnetlink.h>
13 #include <linux/iopoll.h>
14 #include "lan743x_main.h"
16 static void lan743x_pci_cleanup(struct lan743x_adapter *adapter)
18 pci_release_selected_regions(adapter->pdev,
19 pci_select_bars(adapter->pdev,
21 pci_disable_device(adapter->pdev);
24 static int lan743x_pci_init(struct lan743x_adapter *adapter,
27 unsigned long bars = 0;
31 ret = pci_enable_device_mem(pdev);
35 netif_info(adapter, probe, adapter->netdev,
36 "PCI: Vendor ID = 0x%04X, Device ID = 0x%04X\n",
37 pdev->vendor, pdev->device);
38 bars = pci_select_bars(pdev, IORESOURCE_MEM);
39 if (!test_bit(0, &bars))
42 ret = pci_request_selected_regions(pdev, bars, DRIVER_NAME);
50 pci_disable_device(adapter->pdev);
56 static u32 lan743x_csr_read(struct lan743x_adapter *adapter, int offset)
58 return ioread32(&adapter->csr.csr_address[offset]);
61 static void lan743x_csr_write(struct lan743x_adapter *adapter, int offset,
64 iowrite32(data, &adapter->csr.csr_address[offset]);
67 #define LAN743X_CSR_READ_OP(offset) lan743x_csr_read(adapter, offset)
69 static int lan743x_csr_light_reset(struct lan743x_adapter *adapter)
73 data = lan743x_csr_read(adapter, HW_CFG);
75 lan743x_csr_write(adapter, HW_CFG, data);
77 return readx_poll_timeout(LAN743X_CSR_READ_OP, HW_CFG, data,
78 !(data & HW_CFG_LRST_), 100000, 10000000);
81 static int lan743x_csr_wait_for_bit(struct lan743x_adapter *adapter,
82 int offset, u32 bit_mask,
83 int target_value, int usleep_min,
84 int usleep_max, int count)
88 return readx_poll_timeout(LAN743X_CSR_READ_OP, offset, data,
89 target_value == ((data & bit_mask) ? 1 : 0),
90 usleep_max, usleep_min * count);
93 static int lan743x_csr_init(struct lan743x_adapter *adapter)
95 struct lan743x_csr *csr = &adapter->csr;
96 resource_size_t bar_start, bar_length;
99 bar_start = pci_resource_start(adapter->pdev, 0);
100 bar_length = pci_resource_len(adapter->pdev, 0);
101 csr->csr_address = devm_ioremap(&adapter->pdev->dev,
102 bar_start, bar_length);
103 if (!csr->csr_address) {
108 csr->id_rev = lan743x_csr_read(adapter, ID_REV);
109 csr->fpga_rev = lan743x_csr_read(adapter, FPGA_REV);
110 netif_info(adapter, probe, adapter->netdev,
111 "ID_REV = 0x%08X, FPGA_REV = %d.%d\n",
112 csr->id_rev, FPGA_REV_GET_MAJOR_(csr->fpga_rev),
113 FPGA_REV_GET_MINOR_(csr->fpga_rev));
114 if (!ID_REV_IS_VALID_CHIP_ID_(csr->id_rev)) {
119 csr->flags = LAN743X_CSR_FLAG_SUPPORTS_INTR_AUTO_SET_CLR;
120 switch (csr->id_rev & ID_REV_CHIP_REV_MASK_) {
121 case ID_REV_CHIP_REV_A0_:
122 csr->flags |= LAN743X_CSR_FLAG_IS_A0;
123 csr->flags &= ~LAN743X_CSR_FLAG_SUPPORTS_INTR_AUTO_SET_CLR;
125 case ID_REV_CHIP_REV_B0_:
126 csr->flags |= LAN743X_CSR_FLAG_IS_B0;
130 result = lan743x_csr_light_reset(adapter);
138 static void lan743x_intr_software_isr(void *context)
140 struct lan743x_adapter *adapter = context;
141 struct lan743x_intr *intr = &adapter->intr;
144 int_sts = lan743x_csr_read(adapter, INT_STS);
145 if (int_sts & INT_BIT_SW_GP_) {
146 lan743x_csr_write(adapter, INT_STS, INT_BIT_SW_GP_);
147 intr->software_isr_flag = 1;
151 static void lan743x_tx_isr(void *context, u32 int_sts, u32 flags)
153 struct lan743x_tx *tx = context;
154 struct lan743x_adapter *adapter = tx->adapter;
155 bool enable_flag = true;
158 int_en = lan743x_csr_read(adapter, INT_EN_SET);
159 if (flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_CLEAR) {
160 lan743x_csr_write(adapter, INT_EN_CLR,
161 INT_BIT_DMA_TX_(tx->channel_number));
164 if (int_sts & INT_BIT_DMA_TX_(tx->channel_number)) {
165 u32 ioc_bit = DMAC_INT_BIT_TX_IOC_(tx->channel_number);
169 if (flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_READ)
170 dmac_int_sts = lan743x_csr_read(adapter, DMAC_INT_STS);
172 dmac_int_sts = ioc_bit;
173 if (flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_CHECK)
174 dmac_int_en = lan743x_csr_read(adapter,
177 dmac_int_en = ioc_bit;
179 dmac_int_en &= ioc_bit;
180 dmac_int_sts &= dmac_int_en;
181 if (dmac_int_sts & ioc_bit) {
182 napi_schedule(&tx->napi);
183 enable_flag = false;/* poll func will enable later */
189 lan743x_csr_write(adapter, INT_EN_SET,
190 INT_BIT_DMA_TX_(tx->channel_number));
193 static void lan743x_rx_isr(void *context, u32 int_sts, u32 flags)
195 struct lan743x_rx *rx = context;
196 struct lan743x_adapter *adapter = rx->adapter;
197 bool enable_flag = true;
199 if (flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_CLEAR) {
200 lan743x_csr_write(adapter, INT_EN_CLR,
201 INT_BIT_DMA_RX_(rx->channel_number));
204 if (int_sts & INT_BIT_DMA_RX_(rx->channel_number)) {
205 u32 rx_frame_bit = DMAC_INT_BIT_RXFRM_(rx->channel_number);
209 if (flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_READ)
210 dmac_int_sts = lan743x_csr_read(adapter, DMAC_INT_STS);
212 dmac_int_sts = rx_frame_bit;
213 if (flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_CHECK)
214 dmac_int_en = lan743x_csr_read(adapter,
217 dmac_int_en = rx_frame_bit;
219 dmac_int_en &= rx_frame_bit;
220 dmac_int_sts &= dmac_int_en;
221 if (dmac_int_sts & rx_frame_bit) {
222 napi_schedule(&rx->napi);
223 enable_flag = false;/* poll funct will enable later */
229 lan743x_csr_write(adapter, INT_EN_SET,
230 INT_BIT_DMA_RX_(rx->channel_number));
234 static void lan743x_intr_shared_isr(void *context, u32 int_sts, u32 flags)
236 struct lan743x_adapter *adapter = context;
237 unsigned int channel;
239 if (int_sts & INT_BIT_ALL_RX_) {
240 for (channel = 0; channel < LAN743X_USED_RX_CHANNELS;
242 u32 int_bit = INT_BIT_DMA_RX_(channel);
244 if (int_sts & int_bit) {
245 lan743x_rx_isr(&adapter->rx[channel],
251 if (int_sts & INT_BIT_ALL_TX_) {
252 for (channel = 0; channel < LAN743X_USED_TX_CHANNELS;
254 u32 int_bit = INT_BIT_DMA_TX_(channel);
256 if (int_sts & int_bit) {
257 lan743x_tx_isr(&adapter->tx[channel],
263 if (int_sts & INT_BIT_ALL_OTHER_) {
264 if (int_sts & INT_BIT_SW_GP_) {
265 lan743x_intr_software_isr(adapter);
266 int_sts &= ~INT_BIT_SW_GP_;
270 lan743x_csr_write(adapter, INT_EN_CLR, int_sts);
273 static irqreturn_t lan743x_intr_entry_isr(int irq, void *ptr)
275 struct lan743x_vector *vector = ptr;
276 struct lan743x_adapter *adapter = vector->adapter;
277 irqreturn_t result = IRQ_NONE;
281 if (vector->flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_READ) {
282 int_sts = lan743x_csr_read(adapter, INT_STS);
283 } else if (vector->flags &
284 (LAN743X_VECTOR_FLAG_SOURCE_STATUS_R2C |
285 LAN743X_VECTOR_FLAG_SOURCE_ENABLE_R2C)) {
286 int_sts = lan743x_csr_read(adapter, INT_STS_R2C);
288 /* use mask as implied status */
289 int_sts = vector->int_mask | INT_BIT_MAS_;
292 if (!(int_sts & INT_BIT_MAS_))
295 if (vector->flags & LAN743X_VECTOR_FLAG_VECTOR_ENABLE_ISR_CLEAR)
296 /* disable vector interrupt */
297 lan743x_csr_write(adapter,
299 INT_VEC_EN_(vector->vector_index));
301 if (vector->flags & LAN743X_VECTOR_FLAG_MASTER_ENABLE_CLEAR)
302 /* disable master interrupt */
303 lan743x_csr_write(adapter, INT_EN_CLR, INT_BIT_MAS_);
305 if (vector->flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_CHECK) {
306 int_enables = lan743x_csr_read(adapter, INT_EN_SET);
308 /* use vector mask as implied enable mask */
309 int_enables = vector->int_mask;
312 int_sts &= int_enables;
313 int_sts &= vector->int_mask;
315 if (vector->handler) {
316 vector->handler(vector->context,
317 int_sts, vector->flags);
319 /* disable interrupts on this vector */
320 lan743x_csr_write(adapter, INT_EN_CLR,
323 result = IRQ_HANDLED;
326 if (vector->flags & LAN743X_VECTOR_FLAG_MASTER_ENABLE_SET)
327 /* enable master interrupt */
328 lan743x_csr_write(adapter, INT_EN_SET, INT_BIT_MAS_);
330 if (vector->flags & LAN743X_VECTOR_FLAG_VECTOR_ENABLE_ISR_SET)
331 /* enable vector interrupt */
332 lan743x_csr_write(adapter,
334 INT_VEC_EN_(vector->vector_index));
339 static int lan743x_intr_test_isr(struct lan743x_adapter *adapter)
341 struct lan743x_intr *intr = &adapter->intr;
342 int result = -ENODEV;
345 intr->software_isr_flag = 0;
347 /* enable interrupt */
348 lan743x_csr_write(adapter, INT_EN_SET, INT_BIT_SW_GP_);
350 /* activate interrupt here */
351 lan743x_csr_write(adapter, INT_SET, INT_BIT_SW_GP_);
352 while ((timeout > 0) && (!(intr->software_isr_flag))) {
353 usleep_range(1000, 20000);
357 if (intr->software_isr_flag)
360 /* disable interrupts */
361 lan743x_csr_write(adapter, INT_EN_CLR, INT_BIT_SW_GP_);
365 static int lan743x_intr_register_isr(struct lan743x_adapter *adapter,
366 int vector_index, u32 flags,
368 lan743x_vector_handler handler,
371 struct lan743x_vector *vector = &adapter->intr.vector_list
375 vector->adapter = adapter;
376 vector->flags = flags;
377 vector->vector_index = vector_index;
378 vector->int_mask = int_mask;
379 vector->handler = handler;
380 vector->context = context;
382 ret = request_irq(vector->irq,
383 lan743x_intr_entry_isr,
384 (flags & LAN743X_VECTOR_FLAG_IRQ_SHARED) ?
385 IRQF_SHARED : 0, DRIVER_NAME, vector);
387 vector->handler = NULL;
388 vector->context = NULL;
389 vector->int_mask = 0;
395 static void lan743x_intr_unregister_isr(struct lan743x_adapter *adapter,
398 struct lan743x_vector *vector = &adapter->intr.vector_list
401 free_irq(vector->irq, vector);
402 vector->handler = NULL;
403 vector->context = NULL;
404 vector->int_mask = 0;
408 static u32 lan743x_intr_get_vector_flags(struct lan743x_adapter *adapter,
413 for (index = 0; index < LAN743X_MAX_VECTOR_COUNT; index++) {
414 if (adapter->intr.vector_list[index].int_mask & int_mask)
415 return adapter->intr.vector_list[index].flags;
420 static void lan743x_intr_close(struct lan743x_adapter *adapter)
422 struct lan743x_intr *intr = &adapter->intr;
425 lan743x_csr_write(adapter, INT_EN_CLR, INT_BIT_MAS_);
426 lan743x_csr_write(adapter, INT_VEC_EN_CLR, 0x000000FF);
428 for (index = 0; index < LAN743X_MAX_VECTOR_COUNT; index++) {
429 if (intr->flags & INTR_FLAG_IRQ_REQUESTED(index)) {
430 lan743x_intr_unregister_isr(adapter, index);
431 intr->flags &= ~INTR_FLAG_IRQ_REQUESTED(index);
435 if (intr->flags & INTR_FLAG_MSI_ENABLED) {
436 pci_disable_msi(adapter->pdev);
437 intr->flags &= ~INTR_FLAG_MSI_ENABLED;
440 if (intr->flags & INTR_FLAG_MSIX_ENABLED) {
441 pci_disable_msix(adapter->pdev);
442 intr->flags &= ~INTR_FLAG_MSIX_ENABLED;
446 static int lan743x_intr_open(struct lan743x_adapter *adapter)
448 struct msix_entry msix_entries[LAN743X_MAX_VECTOR_COUNT];
449 struct lan743x_intr *intr = &adapter->intr;
450 u32 int_vec_en_auto_clr = 0;
451 u32 int_vec_map0 = 0;
452 u32 int_vec_map1 = 0;
457 intr->number_of_vectors = 0;
459 /* Try to set up MSIX interrupts */
460 memset(&msix_entries[0], 0,
461 sizeof(struct msix_entry) * LAN743X_MAX_VECTOR_COUNT);
462 for (index = 0; index < LAN743X_MAX_VECTOR_COUNT; index++)
463 msix_entries[index].entry = index;
464 ret = pci_enable_msix_range(adapter->pdev,
466 1 + LAN743X_USED_TX_CHANNELS +
467 LAN743X_USED_RX_CHANNELS);
470 intr->flags |= INTR_FLAG_MSIX_ENABLED;
471 intr->number_of_vectors = ret;
472 intr->using_vectors = true;
473 for (index = 0; index < intr->number_of_vectors; index++)
474 intr->vector_list[index].irq = msix_entries
476 netif_info(adapter, ifup, adapter->netdev,
477 "using MSIX interrupts, number of vectors = %d\n",
478 intr->number_of_vectors);
481 /* If MSIX failed try to setup using MSI interrupts */
482 if (!intr->number_of_vectors) {
483 if (!(adapter->csr.flags & LAN743X_CSR_FLAG_IS_A0)) {
484 if (!pci_enable_msi(adapter->pdev)) {
485 intr->flags |= INTR_FLAG_MSI_ENABLED;
486 intr->number_of_vectors = 1;
487 intr->using_vectors = true;
488 intr->vector_list[0].irq =
490 netif_info(adapter, ifup, adapter->netdev,
491 "using MSI interrupts, number of vectors = %d\n",
492 intr->number_of_vectors);
497 /* If MSIX, and MSI failed, setup using legacy interrupt */
498 if (!intr->number_of_vectors) {
499 intr->number_of_vectors = 1;
500 intr->using_vectors = false;
501 intr->vector_list[0].irq = intr->irq;
502 netif_info(adapter, ifup, adapter->netdev,
503 "using legacy interrupts\n");
506 /* At this point we must have at least one irq */
507 lan743x_csr_write(adapter, INT_VEC_EN_CLR, 0xFFFFFFFF);
509 /* map all interrupts to vector 0 */
510 lan743x_csr_write(adapter, INT_VEC_MAP0, 0x00000000);
511 lan743x_csr_write(adapter, INT_VEC_MAP1, 0x00000000);
512 lan743x_csr_write(adapter, INT_VEC_MAP2, 0x00000000);
513 flags = LAN743X_VECTOR_FLAG_SOURCE_STATUS_READ |
514 LAN743X_VECTOR_FLAG_SOURCE_STATUS_W2C |
515 LAN743X_VECTOR_FLAG_SOURCE_ENABLE_CHECK |
516 LAN743X_VECTOR_FLAG_SOURCE_ENABLE_CLEAR;
518 if (intr->using_vectors) {
519 flags |= LAN743X_VECTOR_FLAG_VECTOR_ENABLE_ISR_CLEAR |
520 LAN743X_VECTOR_FLAG_VECTOR_ENABLE_ISR_SET;
522 flags |= LAN743X_VECTOR_FLAG_MASTER_ENABLE_CLEAR |
523 LAN743X_VECTOR_FLAG_MASTER_ENABLE_SET |
524 LAN743X_VECTOR_FLAG_IRQ_SHARED;
527 if (adapter->csr.flags & LAN743X_CSR_FLAG_SUPPORTS_INTR_AUTO_SET_CLR) {
528 flags &= ~LAN743X_VECTOR_FLAG_SOURCE_STATUS_READ;
529 flags &= ~LAN743X_VECTOR_FLAG_SOURCE_STATUS_W2C;
530 flags &= ~LAN743X_VECTOR_FLAG_SOURCE_ENABLE_CLEAR;
531 flags &= ~LAN743X_VECTOR_FLAG_SOURCE_ENABLE_CHECK;
532 flags |= LAN743X_VECTOR_FLAG_SOURCE_STATUS_R2C;
533 flags |= LAN743X_VECTOR_FLAG_SOURCE_ENABLE_R2C;
536 ret = lan743x_intr_register_isr(adapter, 0, flags,
537 INT_BIT_ALL_RX_ | INT_BIT_ALL_TX_ |
539 lan743x_intr_shared_isr, adapter);
542 intr->flags |= INTR_FLAG_IRQ_REQUESTED(0);
544 if (intr->using_vectors)
545 lan743x_csr_write(adapter, INT_VEC_EN_SET,
548 if (!(adapter->csr.flags & LAN743X_CSR_FLAG_IS_A0)) {
549 lan743x_csr_write(adapter, INT_MOD_CFG0, LAN743X_INT_MOD);
550 lan743x_csr_write(adapter, INT_MOD_CFG1, LAN743X_INT_MOD);
551 lan743x_csr_write(adapter, INT_MOD_CFG2, LAN743X_INT_MOD);
552 lan743x_csr_write(adapter, INT_MOD_CFG3, LAN743X_INT_MOD);
553 lan743x_csr_write(adapter, INT_MOD_CFG4, LAN743X_INT_MOD);
554 lan743x_csr_write(adapter, INT_MOD_CFG5, LAN743X_INT_MOD);
555 lan743x_csr_write(adapter, INT_MOD_CFG6, LAN743X_INT_MOD);
556 lan743x_csr_write(adapter, INT_MOD_CFG7, LAN743X_INT_MOD);
557 lan743x_csr_write(adapter, INT_MOD_MAP0, 0x00005432);
558 lan743x_csr_write(adapter, INT_MOD_MAP1, 0x00000001);
559 lan743x_csr_write(adapter, INT_MOD_MAP2, 0x00FFFFFF);
562 /* enable interrupts */
563 lan743x_csr_write(adapter, INT_EN_SET, INT_BIT_MAS_);
564 ret = lan743x_intr_test_isr(adapter);
568 if (intr->number_of_vectors > 1) {
569 int number_of_tx_vectors = intr->number_of_vectors - 1;
571 if (number_of_tx_vectors > LAN743X_USED_TX_CHANNELS)
572 number_of_tx_vectors = LAN743X_USED_TX_CHANNELS;
573 flags = LAN743X_VECTOR_FLAG_SOURCE_STATUS_READ |
574 LAN743X_VECTOR_FLAG_SOURCE_STATUS_W2C |
575 LAN743X_VECTOR_FLAG_SOURCE_ENABLE_CHECK |
576 LAN743X_VECTOR_FLAG_SOURCE_ENABLE_CLEAR |
577 LAN743X_VECTOR_FLAG_VECTOR_ENABLE_ISR_CLEAR |
578 LAN743X_VECTOR_FLAG_VECTOR_ENABLE_ISR_SET;
580 if (adapter->csr.flags &
581 LAN743X_CSR_FLAG_SUPPORTS_INTR_AUTO_SET_CLR) {
582 flags = LAN743X_VECTOR_FLAG_VECTOR_ENABLE_AUTO_CLEAR |
583 LAN743X_VECTOR_FLAG_VECTOR_ENABLE_AUTO_SET |
584 LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_SET |
585 LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_CLEAR |
586 LAN743X_VECTOR_FLAG_SOURCE_STATUS_AUTO_CLEAR;
589 for (index = 0; index < number_of_tx_vectors; index++) {
590 u32 int_bit = INT_BIT_DMA_TX_(index);
591 int vector = index + 1;
593 /* map TX interrupt to vector */
594 int_vec_map1 |= INT_VEC_MAP1_TX_VEC_(index, vector);
595 lan743x_csr_write(adapter, INT_VEC_MAP1, int_vec_map1);
597 LAN743X_VECTOR_FLAG_VECTOR_ENABLE_AUTO_CLEAR) {
598 int_vec_en_auto_clr |= INT_VEC_EN_(vector);
599 lan743x_csr_write(adapter, INT_VEC_EN_AUTO_CLR,
600 int_vec_en_auto_clr);
603 /* Remove TX interrupt from shared mask */
604 intr->vector_list[0].int_mask &= ~int_bit;
605 ret = lan743x_intr_register_isr(adapter, vector, flags,
606 int_bit, lan743x_tx_isr,
607 &adapter->tx[index]);
610 intr->flags |= INTR_FLAG_IRQ_REQUESTED(vector);
612 LAN743X_VECTOR_FLAG_VECTOR_ENABLE_AUTO_SET))
613 lan743x_csr_write(adapter, INT_VEC_EN_SET,
614 INT_VEC_EN_(vector));
617 if ((intr->number_of_vectors - LAN743X_USED_TX_CHANNELS) > 1) {
618 int number_of_rx_vectors = intr->number_of_vectors -
619 LAN743X_USED_TX_CHANNELS - 1;
621 if (number_of_rx_vectors > LAN743X_USED_RX_CHANNELS)
622 number_of_rx_vectors = LAN743X_USED_RX_CHANNELS;
624 flags = LAN743X_VECTOR_FLAG_SOURCE_STATUS_READ |
625 LAN743X_VECTOR_FLAG_SOURCE_STATUS_W2C |
626 LAN743X_VECTOR_FLAG_SOURCE_ENABLE_CHECK |
627 LAN743X_VECTOR_FLAG_SOURCE_ENABLE_CLEAR |
628 LAN743X_VECTOR_FLAG_VECTOR_ENABLE_ISR_CLEAR |
629 LAN743X_VECTOR_FLAG_VECTOR_ENABLE_ISR_SET;
631 if (adapter->csr.flags &
632 LAN743X_CSR_FLAG_SUPPORTS_INTR_AUTO_SET_CLR) {
633 flags = LAN743X_VECTOR_FLAG_VECTOR_ENABLE_AUTO_CLEAR |
634 LAN743X_VECTOR_FLAG_VECTOR_ENABLE_AUTO_SET |
635 LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_SET |
636 LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_CLEAR |
637 LAN743X_VECTOR_FLAG_SOURCE_STATUS_AUTO_CLEAR;
639 for (index = 0; index < number_of_rx_vectors; index++) {
640 int vector = index + 1 + LAN743X_USED_TX_CHANNELS;
641 u32 int_bit = INT_BIT_DMA_RX_(index);
643 /* map RX interrupt to vector */
644 int_vec_map0 |= INT_VEC_MAP0_RX_VEC_(index, vector);
645 lan743x_csr_write(adapter, INT_VEC_MAP0, int_vec_map0);
647 LAN743X_VECTOR_FLAG_VECTOR_ENABLE_AUTO_CLEAR) {
648 int_vec_en_auto_clr |= INT_VEC_EN_(vector);
649 lan743x_csr_write(adapter, INT_VEC_EN_AUTO_CLR,
650 int_vec_en_auto_clr);
653 /* Remove RX interrupt from shared mask */
654 intr->vector_list[0].int_mask &= ~int_bit;
655 ret = lan743x_intr_register_isr(adapter, vector, flags,
656 int_bit, lan743x_rx_isr,
657 &adapter->rx[index]);
660 intr->flags |= INTR_FLAG_IRQ_REQUESTED(vector);
662 lan743x_csr_write(adapter, INT_VEC_EN_SET,
663 INT_VEC_EN_(vector));
669 lan743x_intr_close(adapter);
673 static int lan743x_dp_write(struct lan743x_adapter *adapter,
674 u32 select, u32 addr, u32 length, u32 *buf)
680 mutex_lock(&adapter->dp_lock);
681 if (lan743x_csr_wait_for_bit(adapter, DP_SEL, DP_SEL_DPRDY_,
684 dp_sel = lan743x_csr_read(adapter, DP_SEL);
685 dp_sel &= ~DP_SEL_MASK_;
687 lan743x_csr_write(adapter, DP_SEL, dp_sel);
689 for (i = 0; i < length; i++) {
690 lan743x_csr_write(adapter, DP_ADDR, addr + i);
691 lan743x_csr_write(adapter, DP_DATA_0, buf[i]);
692 lan743x_csr_write(adapter, DP_CMD, DP_CMD_WRITE_);
693 if (lan743x_csr_wait_for_bit(adapter, DP_SEL, DP_SEL_DPRDY_,
700 mutex_unlock(&adapter->dp_lock);
704 static u32 lan743x_mac_mii_access(u16 id, u16 index, int read)
708 ret = (id << MAC_MII_ACC_PHY_ADDR_SHIFT_) &
709 MAC_MII_ACC_PHY_ADDR_MASK_;
710 ret |= (index << MAC_MII_ACC_MIIRINDA_SHIFT_) &
711 MAC_MII_ACC_MIIRINDA_MASK_;
714 ret |= MAC_MII_ACC_MII_READ_;
716 ret |= MAC_MII_ACC_MII_WRITE_;
717 ret |= MAC_MII_ACC_MII_BUSY_;
722 static int lan743x_mac_mii_wait_till_not_busy(struct lan743x_adapter *adapter)
726 return readx_poll_timeout(LAN743X_CSR_READ_OP, MAC_MII_ACC, data,
727 !(data & MAC_MII_ACC_MII_BUSY_), 0, 1000000);
730 static int lan743x_mdiobus_read(struct mii_bus *bus, int phy_id, int index)
732 struct lan743x_adapter *adapter = bus->priv;
736 /* comfirm MII not busy */
737 ret = lan743x_mac_mii_wait_till_not_busy(adapter);
741 /* set the address, index & direction (read from PHY) */
742 mii_access = lan743x_mac_mii_access(phy_id, index, MAC_MII_READ);
743 lan743x_csr_write(adapter, MAC_MII_ACC, mii_access);
744 ret = lan743x_mac_mii_wait_till_not_busy(adapter);
748 val = lan743x_csr_read(adapter, MAC_MII_DATA);
749 return (int)(val & 0xFFFF);
752 static int lan743x_mdiobus_write(struct mii_bus *bus,
753 int phy_id, int index, u16 regval)
755 struct lan743x_adapter *adapter = bus->priv;
759 /* confirm MII not busy */
760 ret = lan743x_mac_mii_wait_till_not_busy(adapter);
764 lan743x_csr_write(adapter, MAC_MII_DATA, val);
766 /* set the address, index & direction (write to PHY) */
767 mii_access = lan743x_mac_mii_access(phy_id, index, MAC_MII_WRITE);
768 lan743x_csr_write(adapter, MAC_MII_ACC, mii_access);
769 ret = lan743x_mac_mii_wait_till_not_busy(adapter);
773 static void lan743x_mac_set_address(struct lan743x_adapter *adapter,
776 u32 addr_lo, addr_hi;
784 lan743x_csr_write(adapter, MAC_RX_ADDRL, addr_lo);
785 lan743x_csr_write(adapter, MAC_RX_ADDRH, addr_hi);
787 ether_addr_copy(adapter->mac_address, addr);
788 netif_info(adapter, drv, adapter->netdev,
789 "MAC address set to %pM\n", addr);
792 static int lan743x_mac_init(struct lan743x_adapter *adapter)
794 bool mac_address_valid = true;
795 struct net_device *netdev;
801 netdev = adapter->netdev;
802 lan743x_csr_write(adapter, MAC_CR, MAC_CR_RST_);
803 ret = lan743x_csr_wait_for_bit(adapter, MAC_CR, MAC_CR_RST_,
804 0, 1000, 20000, 100);
808 /* setup auto duplex, and speed detection */
809 data = lan743x_csr_read(adapter, MAC_CR);
810 data |= MAC_CR_ADD_ | MAC_CR_ASD_;
811 data |= MAC_CR_CNTR_RST_;
812 lan743x_csr_write(adapter, MAC_CR, data);
814 mac_addr_hi = lan743x_csr_read(adapter, MAC_RX_ADDRH);
815 mac_addr_lo = lan743x_csr_read(adapter, MAC_RX_ADDRL);
816 adapter->mac_address[0] = mac_addr_lo & 0xFF;
817 adapter->mac_address[1] = (mac_addr_lo >> 8) & 0xFF;
818 adapter->mac_address[2] = (mac_addr_lo >> 16) & 0xFF;
819 adapter->mac_address[3] = (mac_addr_lo >> 24) & 0xFF;
820 adapter->mac_address[4] = mac_addr_hi & 0xFF;
821 adapter->mac_address[5] = (mac_addr_hi >> 8) & 0xFF;
823 if (((mac_addr_hi & 0x0000FFFF) == 0x0000FFFF) &&
824 mac_addr_lo == 0xFFFFFFFF) {
825 mac_address_valid = false;
826 } else if (!is_valid_ether_addr(adapter->mac_address)) {
827 mac_address_valid = false;
830 if (!mac_address_valid)
831 random_ether_addr(adapter->mac_address);
832 lan743x_mac_set_address(adapter, adapter->mac_address);
833 ether_addr_copy(netdev->dev_addr, adapter->mac_address);
837 static int lan743x_mac_open(struct lan743x_adapter *adapter)
842 temp = lan743x_csr_read(adapter, MAC_RX);
843 lan743x_csr_write(adapter, MAC_RX, temp | MAC_RX_RXEN_);
844 temp = lan743x_csr_read(adapter, MAC_TX);
845 lan743x_csr_write(adapter, MAC_TX, temp | MAC_TX_TXEN_);
849 static void lan743x_mac_close(struct lan743x_adapter *adapter)
853 temp = lan743x_csr_read(adapter, MAC_TX);
854 temp &= ~MAC_TX_TXEN_;
855 lan743x_csr_write(adapter, MAC_TX, temp);
856 lan743x_csr_wait_for_bit(adapter, MAC_TX, MAC_TX_TXD_,
857 1, 1000, 20000, 100);
859 temp = lan743x_csr_read(adapter, MAC_RX);
860 temp &= ~MAC_RX_RXEN_;
861 lan743x_csr_write(adapter, MAC_RX, temp);
862 lan743x_csr_wait_for_bit(adapter, MAC_RX, MAC_RX_RXD_,
863 1, 1000, 20000, 100);
866 static void lan743x_mac_flow_ctrl_set_enables(struct lan743x_adapter *adapter,
867 bool tx_enable, bool rx_enable)
869 u32 flow_setting = 0;
871 /* set maximum pause time because when fifo space frees
872 * up a zero value pause frame will be sent to release the pause
874 flow_setting = MAC_FLOW_CR_FCPT_MASK_;
876 flow_setting |= MAC_FLOW_CR_TX_FCEN_;
878 flow_setting |= MAC_FLOW_CR_RX_FCEN_;
879 lan743x_csr_write(adapter, MAC_FLOW, flow_setting);
882 static int lan743x_mac_set_mtu(struct lan743x_adapter *adapter, int new_mtu)
887 mac_rx = lan743x_csr_read(adapter, MAC_RX);
888 if (mac_rx & MAC_RX_RXEN_) {
890 if (mac_rx & MAC_RX_RXD_) {
891 lan743x_csr_write(adapter, MAC_RX, mac_rx);
892 mac_rx &= ~MAC_RX_RXD_;
894 mac_rx &= ~MAC_RX_RXEN_;
895 lan743x_csr_write(adapter, MAC_RX, mac_rx);
896 lan743x_csr_wait_for_bit(adapter, MAC_RX, MAC_RX_RXD_,
897 1, 1000, 20000, 100);
898 lan743x_csr_write(adapter, MAC_RX, mac_rx | MAC_RX_RXD_);
901 mac_rx &= ~(MAC_RX_MAX_SIZE_MASK_);
902 mac_rx |= (((new_mtu + ETH_HLEN + 4) << MAC_RX_MAX_SIZE_SHIFT_) &
903 MAC_RX_MAX_SIZE_MASK_);
904 lan743x_csr_write(adapter, MAC_RX, mac_rx);
907 mac_rx |= MAC_RX_RXEN_;
908 lan743x_csr_write(adapter, MAC_RX, mac_rx);
914 static int lan743x_phy_reset(struct lan743x_adapter *adapter)
918 /* Only called with in probe, and before mdiobus_register */
920 data = lan743x_csr_read(adapter, PMT_CTL);
921 data |= PMT_CTL_ETH_PHY_RST_;
922 lan743x_csr_write(adapter, PMT_CTL, data);
924 return readx_poll_timeout(LAN743X_CSR_READ_OP, PMT_CTL, data,
925 (!(data & PMT_CTL_ETH_PHY_RST_) &&
926 (data & PMT_CTL_READY_)),
930 static void lan743x_phy_update_flowcontrol(struct lan743x_adapter *adapter,
931 u8 duplex, u16 local_adv,
934 struct lan743x_phy *phy = &adapter->phy;
938 cap = mii_resolve_flowctrl_fdx(local_adv, remote_adv);
940 cap = phy->fc_request_control;
942 lan743x_mac_flow_ctrl_set_enables(adapter,
947 static int lan743x_phy_init(struct lan743x_adapter *adapter)
949 return lan743x_phy_reset(adapter);
952 static void lan743x_phy_link_status_change(struct net_device *netdev)
954 struct lan743x_adapter *adapter = netdev_priv(netdev);
955 struct phy_device *phydev = netdev->phydev;
957 phy_print_status(phydev);
958 if (phydev->state == PHY_RUNNING) {
959 struct ethtool_link_ksettings ksettings;
960 int remote_advertisement = 0;
961 int local_advertisement = 0;
963 memset(&ksettings, 0, sizeof(ksettings));
964 phy_ethtool_get_link_ksettings(netdev, &ksettings);
965 local_advertisement = phy_read(phydev, MII_ADVERTISE);
966 if (local_advertisement < 0)
969 remote_advertisement = phy_read(phydev, MII_LPA);
970 if (remote_advertisement < 0)
973 lan743x_phy_update_flowcontrol(adapter,
974 ksettings.base.duplex,
976 remote_advertisement);
980 static void lan743x_phy_close(struct lan743x_adapter *adapter)
982 struct net_device *netdev = adapter->netdev;
984 phy_stop(netdev->phydev);
985 phy_disconnect(netdev->phydev);
986 netdev->phydev = NULL;
989 static int lan743x_phy_open(struct lan743x_adapter *adapter)
991 struct lan743x_phy *phy = &adapter->phy;
992 struct phy_device *phydev;
993 struct net_device *netdev;
997 netdev = adapter->netdev;
998 phydev = phy_find_first(adapter->mdiobus);
1002 ret = phy_connect_direct(netdev, phydev,
1003 lan743x_phy_link_status_change,
1004 PHY_INTERFACE_MODE_GMII);
1008 /* MAC doesn't support 1000T Half */
1009 phydev->supported &= ~SUPPORTED_1000baseT_Half;
1011 /* support both flow controls */
1012 phy->fc_request_control = (FLOW_CTRL_RX | FLOW_CTRL_TX);
1013 phydev->advertising &= ~(ADVERTISED_Pause | ADVERTISED_Asym_Pause);
1014 mii_adv = (u32)mii_advertise_flowctrl(phy->fc_request_control);
1015 phydev->advertising |= mii_adv_to_ethtool_adv_t(mii_adv);
1016 phy->fc_autoneg = phydev->autoneg;
1019 phy_start_aneg(phydev);
1026 static void lan743x_rfe_update_mac_address(struct lan743x_adapter *adapter)
1029 u32 mac_addr_hi = 0;
1030 u32 mac_addr_lo = 0;
1032 /* Add mac address to perfect Filter */
1033 mac_addr = adapter->mac_address;
1034 mac_addr_lo = ((((u32)(mac_addr[0])) << 0) |
1035 (((u32)(mac_addr[1])) << 8) |
1036 (((u32)(mac_addr[2])) << 16) |
1037 (((u32)(mac_addr[3])) << 24));
1038 mac_addr_hi = ((((u32)(mac_addr[4])) << 0) |
1039 (((u32)(mac_addr[5])) << 8));
1041 lan743x_csr_write(adapter, RFE_ADDR_FILT_LO(0), mac_addr_lo);
1042 lan743x_csr_write(adapter, RFE_ADDR_FILT_HI(0),
1043 mac_addr_hi | RFE_ADDR_FILT_HI_VALID_);
1046 static void lan743x_rfe_set_multicast(struct lan743x_adapter *adapter)
1048 struct net_device *netdev = adapter->netdev;
1049 u32 hash_table[DP_SEL_VHF_HASH_LEN];
1053 rfctl = lan743x_csr_read(adapter, RFE_CTL);
1054 rfctl &= ~(RFE_CTL_AU_ | RFE_CTL_AM_ |
1055 RFE_CTL_DA_PERFECT_ | RFE_CTL_MCAST_HASH_);
1056 rfctl |= RFE_CTL_AB_;
1057 if (netdev->flags & IFF_PROMISC) {
1058 rfctl |= RFE_CTL_AM_ | RFE_CTL_AU_;
1060 if (netdev->flags & IFF_ALLMULTI)
1061 rfctl |= RFE_CTL_AM_;
1064 memset(hash_table, 0, DP_SEL_VHF_HASH_LEN * sizeof(u32));
1065 if (netdev_mc_count(netdev)) {
1066 struct netdev_hw_addr *ha;
1069 rfctl |= RFE_CTL_DA_PERFECT_;
1071 netdev_for_each_mc_addr(ha, netdev) {
1072 /* set first 32 into Perfect Filter */
1074 lan743x_csr_write(adapter,
1075 RFE_ADDR_FILT_HI(i), 0);
1077 data = ha->addr[2] | (data << 8);
1078 data = ha->addr[1] | (data << 8);
1079 data = ha->addr[0] | (data << 8);
1080 lan743x_csr_write(adapter,
1081 RFE_ADDR_FILT_LO(i), data);
1083 data = ha->addr[4] | (data << 8);
1084 data |= RFE_ADDR_FILT_HI_VALID_;
1085 lan743x_csr_write(adapter,
1086 RFE_ADDR_FILT_HI(i), data);
1088 u32 bitnum = (ether_crc(ETH_ALEN, ha->addr) >>
1090 hash_table[bitnum / 32] |= (1 << (bitnum % 32));
1091 rfctl |= RFE_CTL_MCAST_HASH_;
1097 lan743x_dp_write(adapter, DP_SEL_RFE_RAM,
1098 DP_SEL_VHF_VLAN_LEN,
1099 DP_SEL_VHF_HASH_LEN, hash_table);
1100 lan743x_csr_write(adapter, RFE_CTL, rfctl);
1103 static int lan743x_dmac_init(struct lan743x_adapter *adapter)
1107 lan743x_csr_write(adapter, DMAC_CMD, DMAC_CMD_SWR_);
1108 lan743x_csr_wait_for_bit(adapter, DMAC_CMD, DMAC_CMD_SWR_,
1109 0, 1000, 20000, 100);
1110 switch (DEFAULT_DMA_DESCRIPTOR_SPACING) {
1111 case DMA_DESCRIPTOR_SPACING_16:
1112 data = DMAC_CFG_MAX_DSPACE_16_;
1114 case DMA_DESCRIPTOR_SPACING_32:
1115 data = DMAC_CFG_MAX_DSPACE_32_;
1117 case DMA_DESCRIPTOR_SPACING_64:
1118 data = DMAC_CFG_MAX_DSPACE_64_;
1120 case DMA_DESCRIPTOR_SPACING_128:
1121 data = DMAC_CFG_MAX_DSPACE_128_;
1126 if (!(adapter->csr.flags & LAN743X_CSR_FLAG_IS_A0))
1127 data |= DMAC_CFG_COAL_EN_;
1128 data |= DMAC_CFG_CH_ARB_SEL_RX_HIGH_;
1129 data |= DMAC_CFG_MAX_READ_REQ_SET_(6);
1130 lan743x_csr_write(adapter, DMAC_CFG, data);
1131 data = DMAC_COAL_CFG_TIMER_LIMIT_SET_(1);
1132 data |= DMAC_COAL_CFG_TIMER_TX_START_;
1133 data |= DMAC_COAL_CFG_FLUSH_INTS_;
1134 data |= DMAC_COAL_CFG_INT_EXIT_COAL_;
1135 data |= DMAC_COAL_CFG_CSR_EXIT_COAL_;
1136 data |= DMAC_COAL_CFG_TX_THRES_SET_(0x0A);
1137 data |= DMAC_COAL_CFG_RX_THRES_SET_(0x0C);
1138 lan743x_csr_write(adapter, DMAC_COAL_CFG, data);
1139 data = DMAC_OBFF_TX_THRES_SET_(0x08);
1140 data |= DMAC_OBFF_RX_THRES_SET_(0x0A);
1141 lan743x_csr_write(adapter, DMAC_OBFF_CFG, data);
1145 static int lan743x_dmac_tx_get_state(struct lan743x_adapter *adapter,
1150 dmac_cmd = lan743x_csr_read(adapter, DMAC_CMD);
1151 return DMAC_CHANNEL_STATE_SET((dmac_cmd &
1152 DMAC_CMD_START_T_(tx_channel)),
1154 DMAC_CMD_STOP_T_(tx_channel)));
1157 static int lan743x_dmac_tx_wait_till_stopped(struct lan743x_adapter *adapter,
1164 ((result = lan743x_dmac_tx_get_state(adapter, tx_channel)) ==
1165 DMAC_CHANNEL_STATE_STOP_PENDING)) {
1166 usleep_range(1000, 20000);
1169 if (result == DMAC_CHANNEL_STATE_STOP_PENDING)
1174 static int lan743x_dmac_rx_get_state(struct lan743x_adapter *adapter,
1179 dmac_cmd = lan743x_csr_read(adapter, DMAC_CMD);
1180 return DMAC_CHANNEL_STATE_SET((dmac_cmd &
1181 DMAC_CMD_START_R_(rx_channel)),
1183 DMAC_CMD_STOP_R_(rx_channel)));
1186 static int lan743x_dmac_rx_wait_till_stopped(struct lan743x_adapter *adapter,
1193 ((result = lan743x_dmac_rx_get_state(adapter, rx_channel)) ==
1194 DMAC_CHANNEL_STATE_STOP_PENDING)) {
1195 usleep_range(1000, 20000);
1198 if (result == DMAC_CHANNEL_STATE_STOP_PENDING)
1203 static void lan743x_tx_release_desc(struct lan743x_tx *tx,
1204 int descriptor_index, bool cleanup)
1206 struct lan743x_tx_buffer_info *buffer_info = NULL;
1207 struct lan743x_tx_descriptor *descriptor = NULL;
1208 u32 descriptor_type = 0;
1210 descriptor = &tx->ring_cpu_ptr[descriptor_index];
1211 buffer_info = &tx->buffer_info[descriptor_index];
1212 if (!(buffer_info->flags & TX_BUFFER_INFO_FLAG_ACTIVE))
1215 descriptor_type = (descriptor->data0) &
1216 TX_DESC_DATA0_DTYPE_MASK_;
1217 if (descriptor_type == TX_DESC_DATA0_DTYPE_DATA_)
1218 goto clean_up_data_descriptor;
1222 clean_up_data_descriptor:
1223 if (buffer_info->dma_ptr) {
1224 if (buffer_info->flags &
1225 TX_BUFFER_INFO_FLAG_SKB_FRAGMENT) {
1226 dma_unmap_page(&tx->adapter->pdev->dev,
1227 buffer_info->dma_ptr,
1228 buffer_info->buffer_length,
1231 dma_unmap_single(&tx->adapter->pdev->dev,
1232 buffer_info->dma_ptr,
1233 buffer_info->buffer_length,
1236 buffer_info->dma_ptr = 0;
1237 buffer_info->buffer_length = 0;
1239 if (buffer_info->skb) {
1240 dev_kfree_skb(buffer_info->skb);
1241 buffer_info->skb = NULL;
1245 buffer_info->flags &= ~TX_BUFFER_INFO_FLAG_ACTIVE;
1248 memset(buffer_info, 0, sizeof(*buffer_info));
1249 memset(descriptor, 0, sizeof(*descriptor));
1252 static int lan743x_tx_next_index(struct lan743x_tx *tx, int index)
1254 return ((++index) % tx->ring_size);
1257 static void lan743x_tx_release_completed_descriptors(struct lan743x_tx *tx)
1259 while ((*tx->head_cpu_ptr) != (tx->last_head)) {
1260 lan743x_tx_release_desc(tx, tx->last_head, false);
1261 tx->last_head = lan743x_tx_next_index(tx, tx->last_head);
1265 static void lan743x_tx_release_all_descriptors(struct lan743x_tx *tx)
1267 u32 original_head = 0;
1269 original_head = tx->last_head;
1271 lan743x_tx_release_desc(tx, tx->last_head, true);
1272 tx->last_head = lan743x_tx_next_index(tx, tx->last_head);
1273 } while (tx->last_head != original_head);
1274 memset(tx->ring_cpu_ptr, 0,
1275 sizeof(*tx->ring_cpu_ptr) * (tx->ring_size));
1276 memset(tx->buffer_info, 0,
1277 sizeof(*tx->buffer_info) * (tx->ring_size));
1280 static int lan743x_tx_get_desc_cnt(struct lan743x_tx *tx,
1281 struct sk_buff *skb)
1283 int result = 1; /* 1 for the main skb buffer */
1286 if (skb_is_gso(skb))
1287 result++; /* requires an extension descriptor */
1288 nr_frags = skb_shinfo(skb)->nr_frags;
1289 result += nr_frags; /* 1 for each fragment buffer */
1293 static int lan743x_tx_get_avail_desc(struct lan743x_tx *tx)
1295 int last_head = tx->last_head;
1296 int last_tail = tx->last_tail;
1298 if (last_tail >= last_head)
1299 return tx->ring_size - last_tail + last_head - 1;
1301 return last_head - last_tail - 1;
1304 static int lan743x_tx_frame_start(struct lan743x_tx *tx,
1305 unsigned char *first_buffer,
1306 unsigned int first_buffer_length,
1307 unsigned int frame_length,
1310 /* called only from within lan743x_tx_xmit_frame.
1311 * assuming tx->ring_lock has already been acquired.
1313 struct lan743x_tx_descriptor *tx_descriptor = NULL;
1314 struct lan743x_tx_buffer_info *buffer_info = NULL;
1315 struct lan743x_adapter *adapter = tx->adapter;
1316 struct device *dev = &adapter->pdev->dev;
1319 tx->frame_flags |= TX_FRAME_FLAG_IN_PROGRESS;
1320 tx->frame_first = tx->last_tail;
1321 tx->frame_tail = tx->frame_first;
1323 tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
1324 buffer_info = &tx->buffer_info[tx->frame_tail];
1325 dma_ptr = dma_map_single(dev, first_buffer, first_buffer_length,
1327 if (dma_mapping_error(dev, dma_ptr))
1330 tx_descriptor->data1 = DMA_ADDR_LOW32(dma_ptr);
1331 tx_descriptor->data2 = DMA_ADDR_HIGH32(dma_ptr);
1332 tx_descriptor->data3 = (frame_length << 16) &
1333 TX_DESC_DATA3_FRAME_LENGTH_MSS_MASK_;
1335 buffer_info->skb = NULL;
1336 buffer_info->dma_ptr = dma_ptr;
1337 buffer_info->buffer_length = first_buffer_length;
1338 buffer_info->flags |= TX_BUFFER_INFO_FLAG_ACTIVE;
1340 tx->frame_data0 = (first_buffer_length &
1341 TX_DESC_DATA0_BUF_LENGTH_MASK_) |
1342 TX_DESC_DATA0_DTYPE_DATA_ |
1347 tx->frame_data0 |= TX_DESC_DATA0_ICE_ |
1348 TX_DESC_DATA0_IPE_ |
1351 /* data0 will be programmed in one of other frame assembler functions */
1355 static void lan743x_tx_frame_add_lso(struct lan743x_tx *tx,
1356 unsigned int frame_length)
1358 /* called only from within lan743x_tx_xmit_frame.
1359 * assuming tx->ring_lock has already been acquired.
1361 struct lan743x_tx_descriptor *tx_descriptor = NULL;
1362 struct lan743x_tx_buffer_info *buffer_info = NULL;
1364 /* wrap up previous descriptor */
1365 tx->frame_data0 |= TX_DESC_DATA0_EXT_;
1366 tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
1367 tx_descriptor->data0 = tx->frame_data0;
1369 /* move to next descriptor */
1370 tx->frame_tail = lan743x_tx_next_index(tx, tx->frame_tail);
1371 tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
1372 buffer_info = &tx->buffer_info[tx->frame_tail];
1374 /* add extension descriptor */
1375 tx_descriptor->data1 = 0;
1376 tx_descriptor->data2 = 0;
1377 tx_descriptor->data3 = 0;
1379 buffer_info->skb = NULL;
1380 buffer_info->dma_ptr = 0;
1381 buffer_info->buffer_length = 0;
1382 buffer_info->flags |= TX_BUFFER_INFO_FLAG_ACTIVE;
1384 tx->frame_data0 = (frame_length & TX_DESC_DATA0_EXT_PAY_LENGTH_MASK_) |
1385 TX_DESC_DATA0_DTYPE_EXT_ |
1386 TX_DESC_DATA0_EXT_LSO_;
1388 /* data0 will be programmed in one of other frame assembler functions */
1391 static int lan743x_tx_frame_add_fragment(struct lan743x_tx *tx,
1392 const struct skb_frag_struct *fragment,
1393 unsigned int frame_length)
1395 /* called only from within lan743x_tx_xmit_frame
1396 * assuming tx->ring_lock has already been acquired
1398 struct lan743x_tx_descriptor *tx_descriptor = NULL;
1399 struct lan743x_tx_buffer_info *buffer_info = NULL;
1400 struct lan743x_adapter *adapter = tx->adapter;
1401 struct device *dev = &adapter->pdev->dev;
1402 unsigned int fragment_length = 0;
1405 fragment_length = skb_frag_size(fragment);
1406 if (!fragment_length)
1409 /* wrap up previous descriptor */
1410 tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
1411 tx_descriptor->data0 = tx->frame_data0;
1413 /* move to next descriptor */
1414 tx->frame_tail = lan743x_tx_next_index(tx, tx->frame_tail);
1415 tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
1416 buffer_info = &tx->buffer_info[tx->frame_tail];
1417 dma_ptr = skb_frag_dma_map(dev, fragment,
1420 if (dma_mapping_error(dev, dma_ptr)) {
1423 /* cleanup all previously setup descriptors */
1424 desc_index = tx->frame_first;
1425 while (desc_index != tx->frame_tail) {
1426 lan743x_tx_release_desc(tx, desc_index, true);
1427 desc_index = lan743x_tx_next_index(tx, desc_index);
1430 tx->frame_flags &= ~TX_FRAME_FLAG_IN_PROGRESS;
1431 tx->frame_first = 0;
1432 tx->frame_data0 = 0;
1437 tx_descriptor->data1 = DMA_ADDR_LOW32(dma_ptr);
1438 tx_descriptor->data2 = DMA_ADDR_HIGH32(dma_ptr);
1439 tx_descriptor->data3 = (frame_length << 16) &
1440 TX_DESC_DATA3_FRAME_LENGTH_MSS_MASK_;
1442 buffer_info->skb = NULL;
1443 buffer_info->dma_ptr = dma_ptr;
1444 buffer_info->buffer_length = fragment_length;
1445 buffer_info->flags |= TX_BUFFER_INFO_FLAG_ACTIVE;
1446 buffer_info->flags |= TX_BUFFER_INFO_FLAG_SKB_FRAGMENT;
1448 tx->frame_data0 = (fragment_length & TX_DESC_DATA0_BUF_LENGTH_MASK_) |
1449 TX_DESC_DATA0_DTYPE_DATA_ |
1452 /* data0 will be programmed in one of other frame assembler functions */
1456 static void lan743x_tx_frame_end(struct lan743x_tx *tx,
1457 struct sk_buff *skb,
1460 /* called only from within lan743x_tx_xmit_frame
1461 * assuming tx->ring_lock has already been acquired
1463 struct lan743x_tx_descriptor *tx_descriptor = NULL;
1464 struct lan743x_tx_buffer_info *buffer_info = NULL;
1465 struct lan743x_adapter *adapter = tx->adapter;
1466 u32 tx_tail_flags = 0;
1468 /* wrap up previous descriptor */
1469 tx->frame_data0 |= TX_DESC_DATA0_LS_;
1470 tx->frame_data0 |= TX_DESC_DATA0_IOC_;
1472 tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
1473 buffer_info = &tx->buffer_info[tx->frame_tail];
1474 buffer_info->skb = skb;
1476 buffer_info->flags |= TX_BUFFER_INFO_FLAG_IGNORE_SYNC;
1478 tx_descriptor->data0 = tx->frame_data0;
1479 tx->frame_tail = lan743x_tx_next_index(tx, tx->frame_tail);
1480 tx->last_tail = tx->frame_tail;
1484 if (tx->vector_flags & LAN743X_VECTOR_FLAG_VECTOR_ENABLE_AUTO_SET)
1485 tx_tail_flags |= TX_TAIL_SET_TOP_INT_VEC_EN_;
1486 if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_SET)
1487 tx_tail_flags |= TX_TAIL_SET_DMAC_INT_EN_ |
1488 TX_TAIL_SET_TOP_INT_EN_;
1490 lan743x_csr_write(adapter, TX_TAIL(tx->channel_number),
1491 tx_tail_flags | tx->frame_tail);
1492 tx->frame_flags &= ~TX_FRAME_FLAG_IN_PROGRESS;
1495 static netdev_tx_t lan743x_tx_xmit_frame(struct lan743x_tx *tx,
1496 struct sk_buff *skb)
1498 int required_number_of_descriptors = 0;
1499 unsigned int start_frame_length = 0;
1500 unsigned int frame_length = 0;
1501 unsigned int head_length = 0;
1502 unsigned long irq_flags = 0;
1503 bool ignore_sync = false;
1508 required_number_of_descriptors = lan743x_tx_get_desc_cnt(tx, skb);
1510 spin_lock_irqsave(&tx->ring_lock, irq_flags);
1511 if (required_number_of_descriptors >
1512 lan743x_tx_get_avail_desc(tx)) {
1513 if (required_number_of_descriptors > (tx->ring_size - 1)) {
1516 /* save to overflow buffer */
1517 tx->overflow_skb = skb;
1518 netif_stop_queue(tx->adapter->netdev);
1523 /* space available, transmit skb */
1524 head_length = skb_headlen(skb);
1525 frame_length = skb_pagelen(skb);
1526 nr_frags = skb_shinfo(skb)->nr_frags;
1527 start_frame_length = frame_length;
1528 gso = skb_is_gso(skb);
1530 start_frame_length = max(skb_shinfo(skb)->gso_size,
1534 if (lan743x_tx_frame_start(tx,
1535 skb->data, head_length,
1537 skb->ip_summed == CHECKSUM_PARTIAL)) {
1543 lan743x_tx_frame_add_lso(tx, frame_length);
1548 for (j = 0; j < nr_frags; j++) {
1549 const struct skb_frag_struct *frag;
1551 frag = &(skb_shinfo(skb)->frags[j]);
1552 if (lan743x_tx_frame_add_fragment(tx, frag, frame_length)) {
1553 /* upon error no need to call
1554 * lan743x_tx_frame_end
1555 * frame assembler clean up was performed inside
1556 * lan743x_tx_frame_add_fragment
1564 lan743x_tx_frame_end(tx, skb, ignore_sync);
1567 spin_unlock_irqrestore(&tx->ring_lock, irq_flags);
1568 return NETDEV_TX_OK;
1571 static int lan743x_tx_napi_poll(struct napi_struct *napi, int weight)
1573 struct lan743x_tx *tx = container_of(napi, struct lan743x_tx, napi);
1574 struct lan743x_adapter *adapter = tx->adapter;
1575 bool start_transmitter = false;
1576 unsigned long irq_flags = 0;
1580 ioc_bit = DMAC_INT_BIT_TX_IOC_(tx->channel_number);
1581 int_sts = lan743x_csr_read(adapter, DMAC_INT_STS);
1582 if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_W2C)
1583 lan743x_csr_write(adapter, DMAC_INT_STS, ioc_bit);
1584 spin_lock_irqsave(&tx->ring_lock, irq_flags);
1586 /* clean up tx ring */
1587 lan743x_tx_release_completed_descriptors(tx);
1588 if (netif_queue_stopped(adapter->netdev)) {
1589 if (tx->overflow_skb) {
1590 if (lan743x_tx_get_desc_cnt(tx, tx->overflow_skb) <=
1591 lan743x_tx_get_avail_desc(tx))
1592 start_transmitter = true;
1594 netif_wake_queue(adapter->netdev);
1597 spin_unlock_irqrestore(&tx->ring_lock, irq_flags);
1599 if (start_transmitter) {
1600 /* space is now available, transmit overflow skb */
1601 lan743x_tx_xmit_frame(tx, tx->overflow_skb);
1602 tx->overflow_skb = NULL;
1603 netif_wake_queue(adapter->netdev);
1606 if (!napi_complete_done(napi, weight))
1610 lan743x_csr_write(adapter, INT_EN_SET,
1611 INT_BIT_DMA_TX_(tx->channel_number));
1612 lan743x_csr_read(adapter, INT_STS);
1618 static void lan743x_tx_ring_cleanup(struct lan743x_tx *tx)
1620 if (tx->head_cpu_ptr) {
1621 pci_free_consistent(tx->adapter->pdev,
1622 sizeof(*tx->head_cpu_ptr),
1623 (void *)(tx->head_cpu_ptr),
1625 tx->head_cpu_ptr = NULL;
1626 tx->head_dma_ptr = 0;
1628 kfree(tx->buffer_info);
1629 tx->buffer_info = NULL;
1631 if (tx->ring_cpu_ptr) {
1632 pci_free_consistent(tx->adapter->pdev,
1633 tx->ring_allocation_size,
1636 tx->ring_allocation_size = 0;
1637 tx->ring_cpu_ptr = NULL;
1638 tx->ring_dma_ptr = 0;
1643 static int lan743x_tx_ring_init(struct lan743x_tx *tx)
1645 size_t ring_allocation_size = 0;
1646 void *cpu_ptr = NULL;
1650 tx->ring_size = LAN743X_TX_RING_SIZE;
1651 if (tx->ring_size & ~TX_CFG_B_TX_RING_LEN_MASK_) {
1655 ring_allocation_size = ALIGN(tx->ring_size *
1656 sizeof(struct lan743x_tx_descriptor),
1659 cpu_ptr = pci_zalloc_consistent(tx->adapter->pdev,
1660 ring_allocation_size, &dma_ptr);
1666 tx->ring_allocation_size = ring_allocation_size;
1667 tx->ring_cpu_ptr = (struct lan743x_tx_descriptor *)cpu_ptr;
1668 tx->ring_dma_ptr = dma_ptr;
1670 cpu_ptr = kcalloc(tx->ring_size, sizeof(*tx->buffer_info), GFP_KERNEL);
1675 tx->buffer_info = (struct lan743x_tx_buffer_info *)cpu_ptr;
1677 cpu_ptr = pci_zalloc_consistent(tx->adapter->pdev,
1678 sizeof(*tx->head_cpu_ptr), &dma_ptr);
1684 tx->head_cpu_ptr = cpu_ptr;
1685 tx->head_dma_ptr = dma_ptr;
1686 if (tx->head_dma_ptr & 0x3) {
1694 lan743x_tx_ring_cleanup(tx);
1698 static void lan743x_tx_close(struct lan743x_tx *tx)
1700 struct lan743x_adapter *adapter = tx->adapter;
1702 lan743x_csr_write(adapter,
1704 DMAC_CMD_STOP_T_(tx->channel_number));
1705 lan743x_dmac_tx_wait_till_stopped(adapter, tx->channel_number);
1707 lan743x_csr_write(adapter,
1709 DMAC_INT_BIT_TX_IOC_(tx->channel_number));
1710 lan743x_csr_write(adapter, INT_EN_CLR,
1711 INT_BIT_DMA_TX_(tx->channel_number));
1712 napi_disable(&tx->napi);
1713 netif_napi_del(&tx->napi);
1715 lan743x_csr_write(adapter, FCT_TX_CTL,
1716 FCT_TX_CTL_DIS_(tx->channel_number));
1717 lan743x_csr_wait_for_bit(adapter, FCT_TX_CTL,
1718 FCT_TX_CTL_EN_(tx->channel_number),
1719 0, 1000, 20000, 100);
1721 lan743x_tx_release_all_descriptors(tx);
1723 if (tx->overflow_skb) {
1724 dev_kfree_skb(tx->overflow_skb);
1725 tx->overflow_skb = NULL;
1728 lan743x_tx_ring_cleanup(tx);
1731 static int lan743x_tx_open(struct lan743x_tx *tx)
1733 struct lan743x_adapter *adapter = NULL;
1737 adapter = tx->adapter;
1738 ret = lan743x_tx_ring_init(tx);
1742 /* initialize fifo */
1743 lan743x_csr_write(adapter, FCT_TX_CTL,
1744 FCT_TX_CTL_RESET_(tx->channel_number));
1745 lan743x_csr_wait_for_bit(adapter, FCT_TX_CTL,
1746 FCT_TX_CTL_RESET_(tx->channel_number),
1747 0, 1000, 20000, 100);
1750 lan743x_csr_write(adapter, FCT_TX_CTL,
1751 FCT_TX_CTL_EN_(tx->channel_number));
1753 /* reset tx channel */
1754 lan743x_csr_write(adapter, DMAC_CMD,
1755 DMAC_CMD_TX_SWR_(tx->channel_number));
1756 lan743x_csr_wait_for_bit(adapter, DMAC_CMD,
1757 DMAC_CMD_TX_SWR_(tx->channel_number),
1758 0, 1000, 20000, 100);
1760 /* Write TX_BASE_ADDR */
1761 lan743x_csr_write(adapter,
1762 TX_BASE_ADDRH(tx->channel_number),
1763 DMA_ADDR_HIGH32(tx->ring_dma_ptr));
1764 lan743x_csr_write(adapter,
1765 TX_BASE_ADDRL(tx->channel_number),
1766 DMA_ADDR_LOW32(tx->ring_dma_ptr));
1768 /* Write TX_CFG_B */
1769 data = lan743x_csr_read(adapter, TX_CFG_B(tx->channel_number));
1770 data &= ~TX_CFG_B_TX_RING_LEN_MASK_;
1771 data |= ((tx->ring_size) & TX_CFG_B_TX_RING_LEN_MASK_);
1772 if (!(adapter->csr.flags & LAN743X_CSR_FLAG_IS_A0))
1773 data |= TX_CFG_B_TDMABL_512_;
1774 lan743x_csr_write(adapter, TX_CFG_B(tx->channel_number), data);
1776 /* Write TX_CFG_A */
1777 data = TX_CFG_A_TX_TMR_HPWB_SEL_IOC_ | TX_CFG_A_TX_HP_WB_EN_;
1778 if (!(adapter->csr.flags & LAN743X_CSR_FLAG_IS_A0)) {
1779 data |= TX_CFG_A_TX_HP_WB_ON_INT_TMR_;
1780 data |= TX_CFG_A_TX_PF_THRES_SET_(0x10);
1781 data |= TX_CFG_A_TX_PF_PRI_THRES_SET_(0x04);
1782 data |= TX_CFG_A_TX_HP_WB_THRES_SET_(0x07);
1784 lan743x_csr_write(adapter, TX_CFG_A(tx->channel_number), data);
1786 /* Write TX_HEAD_WRITEBACK_ADDR */
1787 lan743x_csr_write(adapter,
1788 TX_HEAD_WRITEBACK_ADDRH(tx->channel_number),
1789 DMA_ADDR_HIGH32(tx->head_dma_ptr));
1790 lan743x_csr_write(adapter,
1791 TX_HEAD_WRITEBACK_ADDRL(tx->channel_number),
1792 DMA_ADDR_LOW32(tx->head_dma_ptr));
1795 tx->last_head = lan743x_csr_read(adapter, TX_HEAD(tx->channel_number));
1799 lan743x_csr_write(adapter, TX_TAIL(tx->channel_number),
1800 (u32)(tx->last_tail));
1801 tx->vector_flags = lan743x_intr_get_vector_flags(adapter,
1803 (tx->channel_number));
1804 netif_napi_add(adapter->netdev,
1805 &tx->napi, lan743x_tx_napi_poll,
1807 napi_enable(&tx->napi);
1810 if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_CLEAR)
1811 data |= TX_CFG_C_TX_TOP_INT_EN_AUTO_CLR_;
1812 if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_AUTO_CLEAR)
1813 data |= TX_CFG_C_TX_DMA_INT_STS_AUTO_CLR_;
1814 if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_R2C)
1815 data |= TX_CFG_C_TX_INT_STS_R2C_MODE_MASK_;
1816 if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_R2C)
1817 data |= TX_CFG_C_TX_INT_EN_R2C_;
1818 lan743x_csr_write(adapter, TX_CFG_C(tx->channel_number), data);
1820 if (!(tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_SET))
1821 lan743x_csr_write(adapter, INT_EN_SET,
1822 INT_BIT_DMA_TX_(tx->channel_number));
1823 lan743x_csr_write(adapter, DMAC_INT_EN_SET,
1824 DMAC_INT_BIT_TX_IOC_(tx->channel_number));
1826 /* start dmac channel */
1827 lan743x_csr_write(adapter, DMAC_CMD,
1828 DMAC_CMD_START_T_(tx->channel_number));
1832 static int lan743x_rx_next_index(struct lan743x_rx *rx, int index)
1834 return ((++index) % rx->ring_size);
1837 static int lan743x_rx_allocate_ring_element(struct lan743x_rx *rx, int index)
1839 struct lan743x_rx_buffer_info *buffer_info;
1840 struct lan743x_rx_descriptor *descriptor;
1843 length = (LAN743X_MAX_FRAME_SIZE + ETH_HLEN + 4 + RX_HEAD_PADDING);
1844 descriptor = &rx->ring_cpu_ptr[index];
1845 buffer_info = &rx->buffer_info[index];
1846 buffer_info->skb = __netdev_alloc_skb(rx->adapter->netdev,
1848 GFP_ATOMIC | GFP_DMA);
1849 if (!(buffer_info->skb))
1851 buffer_info->dma_ptr = dma_map_single(&rx->adapter->pdev->dev,
1852 buffer_info->skb->data,
1855 if (dma_mapping_error(&rx->adapter->pdev->dev,
1856 buffer_info->dma_ptr)) {
1857 buffer_info->dma_ptr = 0;
1861 buffer_info->buffer_length = length;
1862 descriptor->data1 = DMA_ADDR_LOW32(buffer_info->dma_ptr);
1863 descriptor->data2 = DMA_ADDR_HIGH32(buffer_info->dma_ptr);
1864 descriptor->data3 = 0;
1865 descriptor->data0 = (RX_DESC_DATA0_OWN_ |
1866 (length & RX_DESC_DATA0_BUF_LENGTH_MASK_));
1867 skb_reserve(buffer_info->skb, RX_HEAD_PADDING);
1872 static void lan743x_rx_reuse_ring_element(struct lan743x_rx *rx, int index)
1874 struct lan743x_rx_buffer_info *buffer_info;
1875 struct lan743x_rx_descriptor *descriptor;
1877 descriptor = &rx->ring_cpu_ptr[index];
1878 buffer_info = &rx->buffer_info[index];
1880 descriptor->data1 = DMA_ADDR_LOW32(buffer_info->dma_ptr);
1881 descriptor->data2 = DMA_ADDR_HIGH32(buffer_info->dma_ptr);
1882 descriptor->data3 = 0;
1883 descriptor->data0 = (RX_DESC_DATA0_OWN_ |
1884 ((buffer_info->buffer_length) &
1885 RX_DESC_DATA0_BUF_LENGTH_MASK_));
1888 static void lan743x_rx_release_ring_element(struct lan743x_rx *rx, int index)
1890 struct lan743x_rx_buffer_info *buffer_info;
1891 struct lan743x_rx_descriptor *descriptor;
1893 descriptor = &rx->ring_cpu_ptr[index];
1894 buffer_info = &rx->buffer_info[index];
1896 memset(descriptor, 0, sizeof(*descriptor));
1898 if (buffer_info->dma_ptr) {
1899 dma_unmap_single(&rx->adapter->pdev->dev,
1900 buffer_info->dma_ptr,
1901 buffer_info->buffer_length,
1903 buffer_info->dma_ptr = 0;
1906 if (buffer_info->skb) {
1907 dev_kfree_skb(buffer_info->skb);
1908 buffer_info->skb = NULL;
1911 memset(buffer_info, 0, sizeof(*buffer_info));
1914 static int lan743x_rx_process_packet(struct lan743x_rx *rx)
1916 struct skb_shared_hwtstamps *hwtstamps = NULL;
1917 int result = RX_PROCESS_RESULT_NOTHING_TO_DO;
1918 struct lan743x_rx_buffer_info *buffer_info;
1919 struct lan743x_rx_descriptor *descriptor;
1920 int current_head_index = -1;
1921 int extension_index = -1;
1922 int first_index = -1;
1923 int last_index = -1;
1925 current_head_index = *rx->head_cpu_ptr;
1926 if (current_head_index < 0 || current_head_index >= rx->ring_size)
1929 if (rx->last_head < 0 || rx->last_head >= rx->ring_size)
1932 if (rx->last_head != current_head_index) {
1933 descriptor = &rx->ring_cpu_ptr[rx->last_head];
1934 if (descriptor->data0 & RX_DESC_DATA0_OWN_)
1937 if (!(descriptor->data0 & RX_DESC_DATA0_FS_))
1940 first_index = rx->last_head;
1941 if (descriptor->data0 & RX_DESC_DATA0_LS_) {
1942 last_index = rx->last_head;
1946 index = lan743x_rx_next_index(rx, first_index);
1947 while (index != current_head_index) {
1948 descriptor = &rx->ring_cpu_ptr[index];
1949 if (descriptor->data0 & RX_DESC_DATA0_OWN_)
1952 if (descriptor->data0 & RX_DESC_DATA0_LS_) {
1956 index = lan743x_rx_next_index(rx, index);
1959 if (last_index >= 0) {
1960 descriptor = &rx->ring_cpu_ptr[last_index];
1961 if (descriptor->data0 & RX_DESC_DATA0_EXT_) {
1962 /* extension is expected to follow */
1963 int index = lan743x_rx_next_index(rx,
1965 if (index != current_head_index) {
1966 descriptor = &rx->ring_cpu_ptr[index];
1967 if (descriptor->data0 &
1968 RX_DESC_DATA0_OWN_) {
1971 if (descriptor->data0 &
1972 RX_DESC_DATA0_EXT_) {
1973 extension_index = index;
1978 /* extension is not yet available */
1979 /* prevent processing of this packet */
1986 if (first_index >= 0 && last_index >= 0) {
1987 int real_last_index = last_index;
1988 struct sk_buff *skb = NULL;
1992 /* packet is available */
1993 if (first_index == last_index) {
1994 /* single buffer packet */
1997 buffer_info = &rx->buffer_info[first_index];
1998 skb = buffer_info->skb;
1999 descriptor = &rx->ring_cpu_ptr[first_index];
2001 /* unmap from dma */
2002 if (buffer_info->dma_ptr) {
2003 dma_unmap_single(&rx->adapter->pdev->dev,
2004 buffer_info->dma_ptr,
2005 buffer_info->buffer_length,
2007 buffer_info->dma_ptr = 0;
2008 buffer_info->buffer_length = 0;
2010 buffer_info->skb = NULL;
2011 packet_length = RX_DESC_DATA0_FRAME_LENGTH_GET_
2012 (descriptor->data0);
2013 skb_put(skb, packet_length - 4);
2014 skb->protocol = eth_type_trans(skb,
2015 rx->adapter->netdev);
2016 lan743x_rx_allocate_ring_element(rx, first_index);
2018 int index = first_index;
2020 /* multi buffer packet not supported */
2021 /* this should not happen since
2022 * buffers are allocated to be at least jumbo size
2025 /* clean up buffers */
2026 if (first_index <= last_index) {
2027 while ((index >= first_index) &&
2028 (index <= last_index)) {
2029 lan743x_rx_release_ring_element(rx,
2031 lan743x_rx_allocate_ring_element(rx,
2033 index = lan743x_rx_next_index(rx,
2037 while ((index >= first_index) ||
2038 (index <= last_index)) {
2039 lan743x_rx_release_ring_element(rx,
2041 lan743x_rx_allocate_ring_element(rx,
2043 index = lan743x_rx_next_index(rx,
2049 if (extension_index >= 0) {
2050 descriptor = &rx->ring_cpu_ptr[extension_index];
2051 buffer_info = &rx->buffer_info[extension_index];
2053 ts_sec = descriptor->data1;
2054 ts_nsec = (descriptor->data2 &
2055 RX_DESC_DATA2_TS_NS_MASK_);
2056 lan743x_rx_reuse_ring_element(rx, extension_index);
2057 real_last_index = extension_index;
2061 result = RX_PROCESS_RESULT_PACKET_DROPPED;
2065 if (extension_index < 0)
2066 goto pass_packet_to_os;
2067 hwtstamps = skb_hwtstamps(skb);
2069 hwtstamps->hwtstamp = ktime_set(ts_sec, ts_nsec);
2072 /* pass packet to OS */
2073 napi_gro_receive(&rx->napi, skb);
2074 result = RX_PROCESS_RESULT_PACKET_RECEIVED;
2077 /* push tail and head forward */
2078 rx->last_tail = real_last_index;
2079 rx->last_head = lan743x_rx_next_index(rx, real_last_index);
2085 static int lan743x_rx_napi_poll(struct napi_struct *napi, int weight)
2087 struct lan743x_rx *rx = container_of(napi, struct lan743x_rx, napi);
2088 struct lan743x_adapter *adapter = rx->adapter;
2089 u32 rx_tail_flags = 0;
2092 if (rx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_W2C) {
2093 /* clear int status bit before reading packet */
2094 lan743x_csr_write(adapter, DMAC_INT_STS,
2095 DMAC_INT_BIT_RXFRM_(rx->channel_number));
2098 while (count < weight) {
2099 int rx_process_result = -1;
2101 rx_process_result = lan743x_rx_process_packet(rx);
2102 if (rx_process_result == RX_PROCESS_RESULT_PACKET_RECEIVED) {
2104 } else if (rx_process_result ==
2105 RX_PROCESS_RESULT_NOTHING_TO_DO) {
2107 } else if (rx_process_result ==
2108 RX_PROCESS_RESULT_PACKET_DROPPED) {
2112 rx->frame_count += count;
2113 if (count == weight)
2116 if (!napi_complete_done(napi, count))
2119 if (rx->vector_flags & LAN743X_VECTOR_FLAG_VECTOR_ENABLE_AUTO_SET)
2120 rx_tail_flags |= RX_TAIL_SET_TOP_INT_VEC_EN_;
2121 if (rx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_SET) {
2122 rx_tail_flags |= RX_TAIL_SET_TOP_INT_EN_;
2124 lan743x_csr_write(adapter, INT_EN_SET,
2125 INT_BIT_DMA_RX_(rx->channel_number));
2128 /* update RX_TAIL */
2129 lan743x_csr_write(adapter, RX_TAIL(rx->channel_number),
2130 rx_tail_flags | rx->last_tail);
2135 static void lan743x_rx_ring_cleanup(struct lan743x_rx *rx)
2137 if (rx->buffer_info && rx->ring_cpu_ptr) {
2140 for (index = 0; index < rx->ring_size; index++)
2141 lan743x_rx_release_ring_element(rx, index);
2144 if (rx->head_cpu_ptr) {
2145 pci_free_consistent(rx->adapter->pdev,
2146 sizeof(*rx->head_cpu_ptr),
2149 rx->head_cpu_ptr = NULL;
2150 rx->head_dma_ptr = 0;
2153 kfree(rx->buffer_info);
2154 rx->buffer_info = NULL;
2156 if (rx->ring_cpu_ptr) {
2157 pci_free_consistent(rx->adapter->pdev,
2158 rx->ring_allocation_size,
2161 rx->ring_allocation_size = 0;
2162 rx->ring_cpu_ptr = NULL;
2163 rx->ring_dma_ptr = 0;
2170 static int lan743x_rx_ring_init(struct lan743x_rx *rx)
2172 size_t ring_allocation_size = 0;
2173 dma_addr_t dma_ptr = 0;
2174 void *cpu_ptr = NULL;
2178 rx->ring_size = LAN743X_RX_RING_SIZE;
2179 if (rx->ring_size <= 1) {
2183 if (rx->ring_size & ~RX_CFG_B_RX_RING_LEN_MASK_) {
2187 ring_allocation_size = ALIGN(rx->ring_size *
2188 sizeof(struct lan743x_rx_descriptor),
2191 cpu_ptr = pci_zalloc_consistent(rx->adapter->pdev,
2192 ring_allocation_size, &dma_ptr);
2197 rx->ring_allocation_size = ring_allocation_size;
2198 rx->ring_cpu_ptr = (struct lan743x_rx_descriptor *)cpu_ptr;
2199 rx->ring_dma_ptr = dma_ptr;
2201 cpu_ptr = kcalloc(rx->ring_size, sizeof(*rx->buffer_info),
2207 rx->buffer_info = (struct lan743x_rx_buffer_info *)cpu_ptr;
2209 cpu_ptr = pci_zalloc_consistent(rx->adapter->pdev,
2210 sizeof(*rx->head_cpu_ptr), &dma_ptr);
2216 rx->head_cpu_ptr = cpu_ptr;
2217 rx->head_dma_ptr = dma_ptr;
2218 if (rx->head_dma_ptr & 0x3) {
2224 for (index = 0; index < rx->ring_size; index++) {
2225 ret = lan743x_rx_allocate_ring_element(rx, index);
2232 lan743x_rx_ring_cleanup(rx);
2236 static void lan743x_rx_close(struct lan743x_rx *rx)
2238 struct lan743x_adapter *adapter = rx->adapter;
2240 lan743x_csr_write(adapter, FCT_RX_CTL,
2241 FCT_RX_CTL_DIS_(rx->channel_number));
2242 lan743x_csr_wait_for_bit(adapter, FCT_RX_CTL,
2243 FCT_RX_CTL_EN_(rx->channel_number),
2244 0, 1000, 20000, 100);
2246 lan743x_csr_write(adapter, DMAC_CMD,
2247 DMAC_CMD_STOP_R_(rx->channel_number));
2248 lan743x_dmac_rx_wait_till_stopped(adapter, rx->channel_number);
2250 lan743x_csr_write(adapter, DMAC_INT_EN_CLR,
2251 DMAC_INT_BIT_RXFRM_(rx->channel_number));
2252 lan743x_csr_write(adapter, INT_EN_CLR,
2253 INT_BIT_DMA_RX_(rx->channel_number));
2254 napi_disable(&rx->napi);
2256 netif_napi_del(&rx->napi);
2258 lan743x_rx_ring_cleanup(rx);
2261 static int lan743x_rx_open(struct lan743x_rx *rx)
2263 struct lan743x_adapter *adapter = rx->adapter;
2267 rx->frame_count = 0;
2268 ret = lan743x_rx_ring_init(rx);
2272 netif_napi_add(adapter->netdev,
2273 &rx->napi, lan743x_rx_napi_poll,
2276 lan743x_csr_write(adapter, DMAC_CMD,
2277 DMAC_CMD_RX_SWR_(rx->channel_number));
2278 lan743x_csr_wait_for_bit(adapter, DMAC_CMD,
2279 DMAC_CMD_RX_SWR_(rx->channel_number),
2280 0, 1000, 20000, 100);
2282 /* set ring base address */
2283 lan743x_csr_write(adapter,
2284 RX_BASE_ADDRH(rx->channel_number),
2285 DMA_ADDR_HIGH32(rx->ring_dma_ptr));
2286 lan743x_csr_write(adapter,
2287 RX_BASE_ADDRL(rx->channel_number),
2288 DMA_ADDR_LOW32(rx->ring_dma_ptr));
2290 /* set rx write back address */
2291 lan743x_csr_write(adapter,
2292 RX_HEAD_WRITEBACK_ADDRH(rx->channel_number),
2293 DMA_ADDR_HIGH32(rx->head_dma_ptr));
2294 lan743x_csr_write(adapter,
2295 RX_HEAD_WRITEBACK_ADDRL(rx->channel_number),
2296 DMA_ADDR_LOW32(rx->head_dma_ptr));
2297 data = RX_CFG_A_RX_HP_WB_EN_;
2298 if (!(adapter->csr.flags & LAN743X_CSR_FLAG_IS_A0)) {
2299 data |= (RX_CFG_A_RX_WB_ON_INT_TMR_ |
2300 RX_CFG_A_RX_WB_THRES_SET_(0x7) |
2301 RX_CFG_A_RX_PF_THRES_SET_(16) |
2302 RX_CFG_A_RX_PF_PRI_THRES_SET_(4));
2306 lan743x_csr_write(adapter,
2307 RX_CFG_A(rx->channel_number), data);
2310 data = lan743x_csr_read(adapter, RX_CFG_B(rx->channel_number));
2311 data &= ~RX_CFG_B_RX_PAD_MASK_;
2312 if (!RX_HEAD_PADDING)
2313 data |= RX_CFG_B_RX_PAD_0_;
2315 data |= RX_CFG_B_RX_PAD_2_;
2316 data &= ~RX_CFG_B_RX_RING_LEN_MASK_;
2317 data |= ((rx->ring_size) & RX_CFG_B_RX_RING_LEN_MASK_);
2318 data |= RX_CFG_B_TS_ALL_RX_;
2319 if (!(adapter->csr.flags & LAN743X_CSR_FLAG_IS_A0))
2320 data |= RX_CFG_B_RDMABL_512_;
2322 lan743x_csr_write(adapter, RX_CFG_B(rx->channel_number), data);
2323 rx->vector_flags = lan743x_intr_get_vector_flags(adapter,
2325 (rx->channel_number));
2329 if (rx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_CLEAR)
2330 data |= RX_CFG_C_RX_TOP_INT_EN_AUTO_CLR_;
2331 if (rx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_AUTO_CLEAR)
2332 data |= RX_CFG_C_RX_DMA_INT_STS_AUTO_CLR_;
2333 if (rx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_R2C)
2334 data |= RX_CFG_C_RX_INT_STS_R2C_MODE_MASK_;
2335 if (rx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_R2C)
2336 data |= RX_CFG_C_RX_INT_EN_R2C_;
2337 lan743x_csr_write(adapter, RX_CFG_C(rx->channel_number), data);
2339 rx->last_tail = ((u32)(rx->ring_size - 1));
2340 lan743x_csr_write(adapter, RX_TAIL(rx->channel_number),
2342 rx->last_head = lan743x_csr_read(adapter, RX_HEAD(rx->channel_number));
2343 if (rx->last_head) {
2348 napi_enable(&rx->napi);
2350 lan743x_csr_write(adapter, INT_EN_SET,
2351 INT_BIT_DMA_RX_(rx->channel_number));
2352 lan743x_csr_write(adapter, DMAC_INT_STS,
2353 DMAC_INT_BIT_RXFRM_(rx->channel_number));
2354 lan743x_csr_write(adapter, DMAC_INT_EN_SET,
2355 DMAC_INT_BIT_RXFRM_(rx->channel_number));
2356 lan743x_csr_write(adapter, DMAC_CMD,
2357 DMAC_CMD_START_R_(rx->channel_number));
2359 /* initialize fifo */
2360 lan743x_csr_write(adapter, FCT_RX_CTL,
2361 FCT_RX_CTL_RESET_(rx->channel_number));
2362 lan743x_csr_wait_for_bit(adapter, FCT_RX_CTL,
2363 FCT_RX_CTL_RESET_(rx->channel_number),
2364 0, 1000, 20000, 100);
2365 lan743x_csr_write(adapter, FCT_FLOW(rx->channel_number),
2366 FCT_FLOW_CTL_REQ_EN_ |
2367 FCT_FLOW_CTL_ON_THRESHOLD_SET_(0x2A) |
2368 FCT_FLOW_CTL_OFF_THRESHOLD_SET_(0xA));
2371 lan743x_csr_write(adapter, FCT_RX_CTL,
2372 FCT_RX_CTL_EN_(rx->channel_number));
2376 netif_napi_del(&rx->napi);
2377 lan743x_rx_ring_cleanup(rx);
2383 static int lan743x_netdev_close(struct net_device *netdev)
2385 struct lan743x_adapter *adapter = netdev_priv(netdev);
2388 lan743x_tx_close(&adapter->tx[0]);
2390 for (index = 0; index < LAN743X_USED_RX_CHANNELS; index++)
2391 lan743x_rx_close(&adapter->rx[index]);
2393 lan743x_phy_close(adapter);
2395 lan743x_mac_close(adapter);
2397 lan743x_intr_close(adapter);
2402 static int lan743x_netdev_open(struct net_device *netdev)
2404 struct lan743x_adapter *adapter = netdev_priv(netdev);
2408 ret = lan743x_intr_open(adapter);
2412 ret = lan743x_mac_open(adapter);
2416 ret = lan743x_phy_open(adapter);
2420 for (index = 0; index < LAN743X_USED_RX_CHANNELS; index++) {
2421 ret = lan743x_rx_open(&adapter->rx[index]);
2426 ret = lan743x_tx_open(&adapter->tx[0]);
2433 for (index = 0; index < LAN743X_USED_RX_CHANNELS; index++) {
2434 if (adapter->rx[index].ring_cpu_ptr)
2435 lan743x_rx_close(&adapter->rx[index]);
2437 lan743x_phy_close(adapter);
2440 lan743x_mac_close(adapter);
2443 lan743x_intr_close(adapter);
2446 netif_warn(adapter, ifup, adapter->netdev,
2447 "Error opening LAN743x\n");
2451 static netdev_tx_t lan743x_netdev_xmit_frame(struct sk_buff *skb,
2452 struct net_device *netdev)
2454 struct lan743x_adapter *adapter = netdev_priv(netdev);
2456 return lan743x_tx_xmit_frame(&adapter->tx[0], skb);
2459 static int lan743x_netdev_ioctl(struct net_device *netdev,
2460 struct ifreq *ifr, int cmd)
2462 if (!netif_running(netdev))
2464 return phy_mii_ioctl(netdev->phydev, ifr, cmd);
2467 static void lan743x_netdev_set_multicast(struct net_device *netdev)
2469 struct lan743x_adapter *adapter = netdev_priv(netdev);
2471 lan743x_rfe_set_multicast(adapter);
2474 static int lan743x_netdev_change_mtu(struct net_device *netdev, int new_mtu)
2476 struct lan743x_adapter *adapter = netdev_priv(netdev);
2479 ret = lan743x_mac_set_mtu(adapter, new_mtu);
2481 netdev->mtu = new_mtu;
2485 static void lan743x_netdev_get_stats64(struct net_device *netdev,
2486 struct rtnl_link_stats64 *stats)
2488 struct lan743x_adapter *adapter = netdev_priv(netdev);
2490 stats->rx_packets = lan743x_csr_read(adapter, STAT_RX_TOTAL_FRAMES);
2491 stats->tx_packets = lan743x_csr_read(adapter, STAT_TX_TOTAL_FRAMES);
2492 stats->rx_bytes = lan743x_csr_read(adapter,
2493 STAT_RX_UNICAST_BYTE_COUNT) +
2494 lan743x_csr_read(adapter,
2495 STAT_RX_BROADCAST_BYTE_COUNT) +
2496 lan743x_csr_read(adapter,
2497 STAT_RX_MULTICAST_BYTE_COUNT);
2498 stats->tx_bytes = lan743x_csr_read(adapter,
2499 STAT_TX_UNICAST_BYTE_COUNT) +
2500 lan743x_csr_read(adapter,
2501 STAT_TX_BROADCAST_BYTE_COUNT) +
2502 lan743x_csr_read(adapter,
2503 STAT_TX_MULTICAST_BYTE_COUNT);
2504 stats->rx_errors = lan743x_csr_read(adapter, STAT_RX_FCS_ERRORS) +
2505 lan743x_csr_read(adapter,
2506 STAT_RX_ALIGNMENT_ERRORS) +
2507 lan743x_csr_read(adapter, STAT_RX_JABBER_ERRORS) +
2508 lan743x_csr_read(adapter,
2509 STAT_RX_UNDERSIZE_FRAME_ERRORS) +
2510 lan743x_csr_read(adapter,
2511 STAT_RX_OVERSIZE_FRAME_ERRORS);
2512 stats->tx_errors = lan743x_csr_read(adapter, STAT_TX_FCS_ERRORS) +
2513 lan743x_csr_read(adapter,
2514 STAT_TX_EXCESS_DEFERRAL_ERRORS) +
2515 lan743x_csr_read(adapter, STAT_TX_CARRIER_ERRORS);
2516 stats->rx_dropped = lan743x_csr_read(adapter,
2517 STAT_RX_DROPPED_FRAMES);
2518 stats->tx_dropped = lan743x_csr_read(adapter,
2519 STAT_TX_EXCESSIVE_COLLISION);
2520 stats->multicast = lan743x_csr_read(adapter,
2521 STAT_RX_MULTICAST_FRAMES) +
2522 lan743x_csr_read(adapter,
2523 STAT_TX_MULTICAST_FRAMES);
2524 stats->collisions = lan743x_csr_read(adapter,
2525 STAT_TX_SINGLE_COLLISIONS) +
2526 lan743x_csr_read(adapter,
2527 STAT_TX_MULTIPLE_COLLISIONS) +
2528 lan743x_csr_read(adapter,
2529 STAT_TX_LATE_COLLISIONS);
2532 static int lan743x_netdev_set_mac_address(struct net_device *netdev,
2535 struct lan743x_adapter *adapter = netdev_priv(netdev);
2536 struct sockaddr *sock_addr = addr;
2539 ret = eth_prepare_mac_addr_change(netdev, sock_addr);
2542 ether_addr_copy(netdev->dev_addr, sock_addr->sa_data);
2543 lan743x_mac_set_address(adapter, sock_addr->sa_data);
2544 lan743x_rfe_update_mac_address(adapter);
2548 static const struct net_device_ops lan743x_netdev_ops = {
2549 .ndo_open = lan743x_netdev_open,
2550 .ndo_stop = lan743x_netdev_close,
2551 .ndo_start_xmit = lan743x_netdev_xmit_frame,
2552 .ndo_do_ioctl = lan743x_netdev_ioctl,
2553 .ndo_set_rx_mode = lan743x_netdev_set_multicast,
2554 .ndo_change_mtu = lan743x_netdev_change_mtu,
2555 .ndo_get_stats64 = lan743x_netdev_get_stats64,
2556 .ndo_set_mac_address = lan743x_netdev_set_mac_address,
2559 static void lan743x_hardware_cleanup(struct lan743x_adapter *adapter)
2561 lan743x_csr_write(adapter, INT_EN_CLR, 0xFFFFFFFF);
2564 static void lan743x_mdiobus_cleanup(struct lan743x_adapter *adapter)
2566 mdiobus_unregister(adapter->mdiobus);
2569 static void lan743x_full_cleanup(struct lan743x_adapter *adapter)
2571 unregister_netdev(adapter->netdev);
2573 lan743x_mdiobus_cleanup(adapter);
2574 lan743x_hardware_cleanup(adapter);
2575 lan743x_pci_cleanup(adapter);
2578 static int lan743x_hardware_init(struct lan743x_adapter *adapter,
2579 struct pci_dev *pdev)
2581 struct lan743x_tx *tx;
2585 adapter->intr.irq = adapter->pdev->irq;
2586 lan743x_csr_write(adapter, INT_EN_CLR, 0xFFFFFFFF);
2587 mutex_init(&adapter->dp_lock);
2588 ret = lan743x_mac_init(adapter);
2592 ret = lan743x_phy_init(adapter);
2596 lan743x_rfe_update_mac_address(adapter);
2598 ret = lan743x_dmac_init(adapter);
2602 for (index = 0; index < LAN743X_USED_RX_CHANNELS; index++) {
2603 adapter->rx[index].adapter = adapter;
2604 adapter->rx[index].channel_number = index;
2607 tx = &adapter->tx[0];
2608 tx->adapter = adapter;
2609 tx->channel_number = 0;
2610 spin_lock_init(&tx->ring_lock);
2614 static int lan743x_mdiobus_init(struct lan743x_adapter *adapter)
2618 adapter->mdiobus = devm_mdiobus_alloc(&adapter->pdev->dev);
2619 if (!(adapter->mdiobus)) {
2624 adapter->mdiobus->priv = (void *)adapter;
2625 adapter->mdiobus->read = lan743x_mdiobus_read;
2626 adapter->mdiobus->write = lan743x_mdiobus_write;
2627 adapter->mdiobus->name = "lan743x-mdiobus";
2628 snprintf(adapter->mdiobus->id, MII_BUS_ID_SIZE,
2629 "pci-%s", pci_name(adapter->pdev));
2631 /* set to internal PHY id */
2632 adapter->mdiobus->phy_mask = ~(u32)BIT(1);
2634 /* register mdiobus */
2635 ret = mdiobus_register(adapter->mdiobus);
2644 /* lan743x_pcidev_probe - Device Initialization Routine
2645 * @pdev: PCI device information struct
2646 * @id: entry in lan743x_pci_tbl
2648 * Returns 0 on success, negative on failure
2650 * initializes an adapter identified by a pci_dev structure.
2651 * The OS initialization, configuring of the adapter private structure,
2652 * and a hardware reset occur.
2654 static int lan743x_pcidev_probe(struct pci_dev *pdev,
2655 const struct pci_device_id *id)
2657 struct lan743x_adapter *adapter = NULL;
2658 struct net_device *netdev = NULL;
2661 netdev = devm_alloc_etherdev(&pdev->dev,
2662 sizeof(struct lan743x_adapter));
2666 SET_NETDEV_DEV(netdev, &pdev->dev);
2667 pci_set_drvdata(pdev, netdev);
2668 adapter = netdev_priv(netdev);
2669 adapter->netdev = netdev;
2670 adapter->msg_enable = NETIF_MSG_DRV | NETIF_MSG_PROBE |
2671 NETIF_MSG_LINK | NETIF_MSG_IFUP |
2672 NETIF_MSG_IFDOWN | NETIF_MSG_TX_QUEUED;
2673 netdev->max_mtu = LAN743X_MAX_FRAME_SIZE;
2675 ret = lan743x_pci_init(adapter, pdev);
2679 ret = lan743x_csr_init(adapter);
2683 ret = lan743x_hardware_init(adapter, pdev);
2687 ret = lan743x_mdiobus_init(adapter);
2689 goto cleanup_hardware;
2691 adapter->netdev->netdev_ops = &lan743x_netdev_ops;
2692 adapter->netdev->features = NETIF_F_SG | NETIF_F_TSO | NETIF_F_HW_CSUM;
2693 adapter->netdev->hw_features = adapter->netdev->features;
2695 /* carrier off reporting is important to ethtool even BEFORE open */
2696 netif_carrier_off(netdev);
2698 ret = register_netdev(adapter->netdev);
2700 goto cleanup_mdiobus;
2704 lan743x_mdiobus_cleanup(adapter);
2707 lan743x_hardware_cleanup(adapter);
2710 lan743x_pci_cleanup(adapter);
2713 pr_warn("Initialization failed\n");
2718 * lan743x_pcidev_remove - Device Removal Routine
2719 * @pdev: PCI device information struct
2721 * this is called by the PCI subsystem to alert the driver
2722 * that it should release a PCI device. This could be caused by a
2723 * Hot-Plug event, or because the driver is going to be removed from
2726 static void lan743x_pcidev_remove(struct pci_dev *pdev)
2728 struct net_device *netdev = pci_get_drvdata(pdev);
2729 struct lan743x_adapter *adapter = netdev_priv(netdev);
2731 lan743x_full_cleanup(adapter);
2734 static void lan743x_pcidev_shutdown(struct pci_dev *pdev)
2736 struct net_device *netdev = pci_get_drvdata(pdev);
2737 struct lan743x_adapter *adapter = netdev_priv(netdev);
2740 netif_device_detach(netdev);
2742 /* close netdev when netdev is at running state.
2743 * For instance, it is true when system goes to sleep by pm-suspend
2744 * However, it is false when system goes to sleep by suspend GUI menu
2746 if (netif_running(netdev))
2747 lan743x_netdev_close(netdev);
2750 /* clean up lan743x portion */
2751 lan743x_hardware_cleanup(adapter);
2754 static const struct pci_device_id lan743x_pcidev_tbl[] = {
2755 { PCI_DEVICE(PCI_VENDOR_ID_SMSC, PCI_DEVICE_ID_SMSC_LAN7430) },
2759 static struct pci_driver lan743x_pcidev_driver = {
2760 .name = DRIVER_NAME,
2761 .id_table = lan743x_pcidev_tbl,
2762 .probe = lan743x_pcidev_probe,
2763 .remove = lan743x_pcidev_remove,
2764 .shutdown = lan743x_pcidev_shutdown,
2767 module_pci_driver(lan743x_pcidev_driver);
2769 MODULE_AUTHOR(DRIVER_AUTHOR);
2770 MODULE_DESCRIPTION(DRIVER_DESC);
2771 MODULE_LICENSE("GPL");