1 // SPDX-License-Identifier: GPL-2.0-only
3 * FUJITSU Extended Socket Network Device driver
4 * Copyright (c) 2015 FUJITSU LIMITED
7 #include <linux/module.h>
8 #include <linux/types.h>
10 #include <linux/platform_device.h>
11 #include <linux/netdevice.h>
12 #include <linux/interrupt.h>
15 #include "fjes_trace.h"
19 #define DRV_VERSION __stringify(MAJ) "." __stringify(MIN)
20 #define DRV_NAME "fjes"
21 char fjes_driver_name[] = DRV_NAME;
22 char fjes_driver_version[] = DRV_VERSION;
23 static const char fjes_driver_string[] =
24 "FUJITSU Extended Socket Network Device Driver";
25 static const char fjes_copyright[] =
26 "Copyright (c) 2015 FUJITSU LIMITED";
29 MODULE_DESCRIPTION("FUJITSU Extended Socket Network Device Driver");
30 MODULE_LICENSE("GPL");
31 MODULE_VERSION(DRV_VERSION);
33 #define ACPI_MOTHERBOARD_RESOURCE_HID "PNP0C02"
35 static const struct acpi_device_id fjes_acpi_ids[] = {
36 {ACPI_MOTHERBOARD_RESOURCE_HID, 0},
39 MODULE_DEVICE_TABLE(acpi, fjes_acpi_ids);
41 static bool is_extended_socket_device(struct acpi_device *device)
43 struct acpi_buffer buffer = { ACPI_ALLOCATE_BUFFER, NULL};
44 char str_buf[sizeof(FJES_ACPI_SYMBOL) + 1];
45 union acpi_object *str;
49 status = acpi_evaluate_object(device->handle, "_STR", NULL, &buffer);
50 if (ACPI_FAILURE(status))
54 result = utf16s_to_utf8s((wchar_t *)str->string.pointer,
55 str->string.length, UTF16_LITTLE_ENDIAN,
56 str_buf, sizeof(str_buf) - 1);
59 if (strncmp(FJES_ACPI_SYMBOL, str_buf, strlen(FJES_ACPI_SYMBOL)) != 0) {
60 kfree(buffer.pointer);
63 kfree(buffer.pointer);
68 static int acpi_check_extended_socket_status(struct acpi_device *device)
70 unsigned long long sta;
73 status = acpi_evaluate_integer(device->handle, "_STA", NULL, &sta);
74 if (ACPI_FAILURE(status))
77 if (!((sta & ACPI_STA_DEVICE_PRESENT) &&
78 (sta & ACPI_STA_DEVICE_ENABLED) &&
79 (sta & ACPI_STA_DEVICE_UI) &&
80 (sta & ACPI_STA_DEVICE_FUNCTIONING)))
87 fjes_get_acpi_resource(struct acpi_resource *acpi_res, void *data)
89 struct acpi_resource_address32 *addr;
90 struct acpi_resource_irq *irq;
91 struct resource *res = data;
93 switch (acpi_res->type) {
94 case ACPI_RESOURCE_TYPE_ADDRESS32:
95 addr = &acpi_res->data.address32;
96 res[0].start = addr->address.minimum;
97 res[0].end = addr->address.minimum +
98 addr->address.address_length - 1;
101 case ACPI_RESOURCE_TYPE_IRQ:
102 irq = &acpi_res->data.irq;
103 if (irq->interrupt_count != 1)
105 res[1].start = irq->interrupts[0];
106 res[1].end = irq->interrupts[0];
116 static struct resource fjes_resource[] = {
117 DEFINE_RES_MEM(0, 1),
121 static int fjes_acpi_add(struct acpi_device *device)
123 struct platform_device *plat_dev;
126 if (!is_extended_socket_device(device))
129 if (acpi_check_extended_socket_status(device))
132 status = acpi_walk_resources(device->handle, METHOD_NAME__CRS,
133 fjes_get_acpi_resource, fjes_resource);
134 if (ACPI_FAILURE(status))
137 /* create platform_device */
138 plat_dev = platform_device_register_simple(DRV_NAME, 0, fjes_resource,
139 ARRAY_SIZE(fjes_resource));
140 if (IS_ERR(plat_dev))
141 return PTR_ERR(plat_dev);
143 device->driver_data = plat_dev;
148 static void fjes_acpi_remove(struct acpi_device *device)
150 struct platform_device *plat_dev;
152 plat_dev = (struct platform_device *)acpi_driver_data(device);
153 platform_device_unregister(plat_dev);
156 static struct acpi_driver fjes_acpi_driver = {
159 .owner = THIS_MODULE,
160 .ids = fjes_acpi_ids,
162 .add = fjes_acpi_add,
163 .remove = fjes_acpi_remove,
167 static int fjes_setup_resources(struct fjes_adapter *adapter)
169 struct net_device *netdev = adapter->netdev;
170 struct ep_share_mem_info *buf_pair;
171 struct fjes_hw *hw = &adapter->hw;
176 mutex_lock(&hw->hw_info.lock);
177 result = fjes_hw_request_info(hw);
180 for (epidx = 0; epidx < hw->max_epid; epidx++) {
181 hw->ep_shm_info[epidx].es_status =
182 hw->hw_info.res_buf->info.info[epidx].es_status;
183 hw->ep_shm_info[epidx].zone =
184 hw->hw_info.res_buf->info.info[epidx].zone;
190 adapter->force_reset = true;
192 mutex_unlock(&hw->hw_info.lock);
195 mutex_unlock(&hw->hw_info.lock);
197 for (epidx = 0; epidx < (hw->max_epid); epidx++) {
198 if ((epidx != hw->my_epid) &&
199 (hw->ep_shm_info[epidx].es_status ==
200 FJES_ZONING_STATUS_ENABLE)) {
201 fjes_hw_raise_interrupt(hw, epidx,
202 REG_ICTL_MASK_INFO_UPDATE);
203 hw->ep_shm_info[epidx].ep_stats
204 .send_intr_zoneupdate += 1;
208 msleep(FJES_OPEN_ZONE_UPDATE_WAIT * hw->max_epid);
210 for (epidx = 0; epidx < (hw->max_epid); epidx++) {
211 if (epidx == hw->my_epid)
214 buf_pair = &hw->ep_shm_info[epidx];
216 spin_lock_irqsave(&hw->rx_status_lock, flags);
217 fjes_hw_setup_epbuf(&buf_pair->tx, netdev->dev_addr,
219 spin_unlock_irqrestore(&hw->rx_status_lock, flags);
221 if (fjes_hw_epid_is_same_zone(hw, epidx)) {
222 mutex_lock(&hw->hw_info.lock);
224 fjes_hw_register_buff_addr(hw, epidx, buf_pair);
225 mutex_unlock(&hw->hw_info.lock);
233 adapter->force_reset = true;
237 hw->ep_shm_info[epidx].ep_stats
238 .com_regist_buf_exec += 1;
245 static void fjes_rx_irq(struct fjes_adapter *adapter, int src_epid)
247 struct fjes_hw *hw = &adapter->hw;
249 fjes_hw_set_irqmask(hw, REG_ICTL_MASK_RX_DATA, true);
251 adapter->unset_rx_last = true;
252 napi_schedule(&adapter->napi);
255 static void fjes_stop_req_irq(struct fjes_adapter *adapter, int src_epid)
257 struct fjes_hw *hw = &adapter->hw;
258 enum ep_partner_status status;
261 set_bit(src_epid, &hw->hw_info.buffer_unshare_reserve_bit);
263 status = fjes_hw_get_partner_ep_status(hw, src_epid);
264 trace_fjes_stop_req_irq_pre(hw, src_epid, status);
266 case EP_PARTNER_WAITING:
267 spin_lock_irqsave(&hw->rx_status_lock, flags);
268 hw->ep_shm_info[src_epid].tx.info->v1i.rx_status |=
269 FJES_RX_STOP_REQ_DONE;
270 spin_unlock_irqrestore(&hw->rx_status_lock, flags);
271 clear_bit(src_epid, &hw->txrx_stop_req_bit);
273 case EP_PARTNER_UNSHARE:
274 case EP_PARTNER_COMPLETE:
276 set_bit(src_epid, &adapter->unshare_watch_bitmask);
277 if (!work_pending(&adapter->unshare_watch_task))
278 queue_work(adapter->control_wq,
279 &adapter->unshare_watch_task);
281 case EP_PARTNER_SHARED:
282 set_bit(src_epid, &hw->epstop_req_bit);
284 if (!work_pending(&hw->epstop_task))
285 queue_work(adapter->control_wq, &hw->epstop_task);
288 trace_fjes_stop_req_irq_post(hw, src_epid);
291 static void fjes_txrx_stop_req_irq(struct fjes_adapter *adapter,
294 struct fjes_hw *hw = &adapter->hw;
295 enum ep_partner_status status;
298 status = fjes_hw_get_partner_ep_status(hw, src_epid);
299 trace_fjes_txrx_stop_req_irq_pre(hw, src_epid, status);
301 case EP_PARTNER_UNSHARE:
302 case EP_PARTNER_COMPLETE:
305 case EP_PARTNER_WAITING:
306 if (src_epid < hw->my_epid) {
307 spin_lock_irqsave(&hw->rx_status_lock, flags);
308 hw->ep_shm_info[src_epid].tx.info->v1i.rx_status |=
309 FJES_RX_STOP_REQ_DONE;
310 spin_unlock_irqrestore(&hw->rx_status_lock, flags);
312 clear_bit(src_epid, &hw->txrx_stop_req_bit);
313 set_bit(src_epid, &adapter->unshare_watch_bitmask);
315 if (!work_pending(&adapter->unshare_watch_task))
316 queue_work(adapter->control_wq,
317 &adapter->unshare_watch_task);
320 case EP_PARTNER_SHARED:
321 if (hw->ep_shm_info[src_epid].rx.info->v1i.rx_status &
322 FJES_RX_STOP_REQ_REQUEST) {
323 set_bit(src_epid, &hw->epstop_req_bit);
324 if (!work_pending(&hw->epstop_task))
325 queue_work(adapter->control_wq,
330 trace_fjes_txrx_stop_req_irq_post(hw, src_epid);
333 static void fjes_update_zone_irq(struct fjes_adapter *adapter,
336 struct fjes_hw *hw = &adapter->hw;
338 if (!work_pending(&hw->update_zone_task))
339 queue_work(adapter->control_wq, &hw->update_zone_task);
342 static irqreturn_t fjes_intr(int irq, void *data)
344 struct fjes_adapter *adapter = data;
345 struct fjes_hw *hw = &adapter->hw;
349 icr = fjes_hw_capture_interrupt_status(hw);
351 if (icr & REG_IS_MASK_IS_ASSERT) {
352 if (icr & REG_ICTL_MASK_RX_DATA) {
353 fjes_rx_irq(adapter, icr & REG_IS_MASK_EPID);
354 hw->ep_shm_info[icr & REG_IS_MASK_EPID].ep_stats
358 if (icr & REG_ICTL_MASK_DEV_STOP_REQ) {
359 fjes_stop_req_irq(adapter, icr & REG_IS_MASK_EPID);
360 hw->ep_shm_info[icr & REG_IS_MASK_EPID].ep_stats
361 .recv_intr_stop += 1;
364 if (icr & REG_ICTL_MASK_TXRX_STOP_REQ) {
365 fjes_txrx_stop_req_irq(adapter, icr & REG_IS_MASK_EPID);
366 hw->ep_shm_info[icr & REG_IS_MASK_EPID].ep_stats
367 .recv_intr_unshare += 1;
370 if (icr & REG_ICTL_MASK_TXRX_STOP_DONE)
371 fjes_hw_set_irqmask(hw,
372 REG_ICTL_MASK_TXRX_STOP_DONE, true);
374 if (icr & REG_ICTL_MASK_INFO_UPDATE) {
375 fjes_update_zone_irq(adapter, icr & REG_IS_MASK_EPID);
376 hw->ep_shm_info[icr & REG_IS_MASK_EPID].ep_stats
377 .recv_intr_zoneupdate += 1;
388 static int fjes_request_irq(struct fjes_adapter *adapter)
390 struct net_device *netdev = adapter->netdev;
393 adapter->interrupt_watch_enable = true;
394 if (!delayed_work_pending(&adapter->interrupt_watch_task)) {
395 queue_delayed_work(adapter->control_wq,
396 &adapter->interrupt_watch_task,
397 FJES_IRQ_WATCH_DELAY);
400 if (!adapter->irq_registered) {
401 result = request_irq(adapter->hw.hw_res.irq, fjes_intr,
402 IRQF_SHARED, netdev->name, adapter);
404 adapter->irq_registered = false;
406 adapter->irq_registered = true;
412 static void fjes_free_irq(struct fjes_adapter *adapter)
414 struct fjes_hw *hw = &adapter->hw;
416 adapter->interrupt_watch_enable = false;
417 cancel_delayed_work_sync(&adapter->interrupt_watch_task);
419 fjes_hw_set_irqmask(hw, REG_ICTL_MASK_ALL, true);
421 if (adapter->irq_registered) {
422 free_irq(adapter->hw.hw_res.irq, adapter);
423 adapter->irq_registered = false;
427 static void fjes_free_resources(struct fjes_adapter *adapter)
429 struct net_device *netdev = adapter->netdev;
430 struct fjes_device_command_param param;
431 struct ep_share_mem_info *buf_pair;
432 struct fjes_hw *hw = &adapter->hw;
433 bool reset_flag = false;
438 for (epidx = 0; epidx < hw->max_epid; epidx++) {
439 if (epidx == hw->my_epid)
442 mutex_lock(&hw->hw_info.lock);
443 result = fjes_hw_unregister_buff_addr(hw, epidx);
444 mutex_unlock(&hw->hw_info.lock);
446 hw->ep_shm_info[epidx].ep_stats.com_unregist_buf_exec += 1;
451 buf_pair = &hw->ep_shm_info[epidx];
453 spin_lock_irqsave(&hw->rx_status_lock, flags);
454 fjes_hw_setup_epbuf(&buf_pair->tx,
455 netdev->dev_addr, netdev->mtu);
456 spin_unlock_irqrestore(&hw->rx_status_lock, flags);
458 clear_bit(epidx, &hw->txrx_stop_req_bit);
461 if (reset_flag || adapter->force_reset) {
462 result = fjes_hw_reset(hw);
464 adapter->force_reset = false;
467 adapter->open_guard = true;
469 hw->hw_info.buffer_share_bit = 0;
471 memset((void *)¶m, 0, sizeof(param));
473 param.req_len = hw->hw_info.req_buf_size;
474 param.req_start = __pa(hw->hw_info.req_buf);
475 param.res_len = hw->hw_info.res_buf_size;
476 param.res_start = __pa(hw->hw_info.res_buf);
477 param.share_start = __pa(hw->hw_info.share->ep_status);
479 fjes_hw_init_command_registers(hw, ¶m);
483 /* fjes_open - Called when a network interface is made active */
484 static int fjes_open(struct net_device *netdev)
486 struct fjes_adapter *adapter = netdev_priv(netdev);
487 struct fjes_hw *hw = &adapter->hw;
490 if (adapter->open_guard)
493 result = fjes_setup_resources(adapter);
497 hw->txrx_stop_req_bit = 0;
498 hw->epstop_req_bit = 0;
500 napi_enable(&adapter->napi);
502 fjes_hw_capture_interrupt_status(hw);
504 result = fjes_request_irq(adapter);
508 fjes_hw_set_irqmask(hw, REG_ICTL_MASK_ALL, false);
510 netif_tx_start_all_queues(netdev);
511 netif_carrier_on(netdev);
516 fjes_free_irq(adapter);
517 napi_disable(&adapter->napi);
520 fjes_free_resources(adapter);
524 /* fjes_close - Disables a network interface */
525 static int fjes_close(struct net_device *netdev)
527 struct fjes_adapter *adapter = netdev_priv(netdev);
528 struct fjes_hw *hw = &adapter->hw;
532 netif_tx_stop_all_queues(netdev);
533 netif_carrier_off(netdev);
535 fjes_hw_raise_epstop(hw);
537 napi_disable(&adapter->napi);
539 spin_lock_irqsave(&hw->rx_status_lock, flags);
540 for (epidx = 0; epidx < hw->max_epid; epidx++) {
541 if (epidx == hw->my_epid)
544 if (fjes_hw_get_partner_ep_status(hw, epidx) ==
546 adapter->hw.ep_shm_info[epidx]
547 .tx.info->v1i.rx_status &=
550 spin_unlock_irqrestore(&hw->rx_status_lock, flags);
552 fjes_free_irq(adapter);
554 cancel_delayed_work_sync(&adapter->interrupt_watch_task);
555 cancel_work_sync(&adapter->unshare_watch_task);
556 adapter->unshare_watch_bitmask = 0;
557 cancel_work_sync(&adapter->raise_intr_rxdata_task);
558 cancel_work_sync(&adapter->tx_stall_task);
560 cancel_work_sync(&hw->update_zone_task);
561 cancel_work_sync(&hw->epstop_task);
563 fjes_hw_wait_epstop(hw);
565 fjes_free_resources(adapter);
570 static int fjes_tx_send(struct fjes_adapter *adapter, int dest,
571 void *data, size_t len)
575 retval = fjes_hw_epbuf_tx_pkt_send(&adapter->hw.ep_shm_info[dest].tx,
580 adapter->hw.ep_shm_info[dest].tx.info->v1i.tx_status =
581 FJES_TX_DELAY_SEND_PENDING;
582 if (!work_pending(&adapter->raise_intr_rxdata_task))
583 queue_work(adapter->txrx_wq,
584 &adapter->raise_intr_rxdata_task);
591 fjes_xmit_frame(struct sk_buff *skb, struct net_device *netdev)
593 struct fjes_adapter *adapter = netdev_priv(netdev);
594 struct fjes_hw *hw = &adapter->hw;
596 int max_epid, my_epid, dest_epid;
597 enum ep_partner_status pstatus;
598 struct netdev_queue *cur_queue;
599 char shortpkt[VLAN_ETH_HLEN];
610 cur_queue = netdev_get_tx_queue(netdev, queue_no);
612 eth = (struct ethhdr *)skb->data;
613 my_epid = hw->my_epid;
615 vlan = (vlan_get_tag(skb, &vlan_id) == 0) ? true : false;
620 if (is_multicast_ether_addr(eth->h_dest)) {
622 max_epid = hw->max_epid;
624 } else if (is_local_ether_addr(eth->h_dest)) {
625 dest_epid = eth->h_dest[ETH_ALEN - 1];
626 max_epid = dest_epid + 1;
628 if ((eth->h_dest[0] == 0x02) &&
629 (0x00 == (eth->h_dest[1] | eth->h_dest[2] |
630 eth->h_dest[3] | eth->h_dest[4])) &&
631 (dest_epid < hw->max_epid)) {
638 adapter->stats64.tx_packets += 1;
639 hw->ep_shm_info[my_epid].net_stats.tx_packets += 1;
640 adapter->stats64.tx_bytes += len;
641 hw->ep_shm_info[my_epid].net_stats.tx_bytes += len;
648 adapter->stats64.tx_packets += 1;
649 hw->ep_shm_info[my_epid].net_stats.tx_packets += 1;
650 adapter->stats64.tx_bytes += len;
651 hw->ep_shm_info[my_epid].net_stats.tx_bytes += len;
654 for (; dest_epid < max_epid; dest_epid++) {
655 if (my_epid == dest_epid)
658 pstatus = fjes_hw_get_partner_ep_status(hw, dest_epid);
659 if (pstatus != EP_PARTNER_SHARED) {
661 hw->ep_shm_info[dest_epid].ep_stats
662 .tx_dropped_not_shared += 1;
664 } else if (!fjes_hw_check_epbuf_version(
665 &adapter->hw.ep_shm_info[dest_epid].rx, 0)) {
666 /* version is NOT 0 */
667 adapter->stats64.tx_carrier_errors += 1;
668 hw->ep_shm_info[dest_epid].net_stats
669 .tx_carrier_errors += 1;
670 hw->ep_shm_info[dest_epid].ep_stats
671 .tx_dropped_ver_mismatch += 1;
674 } else if (!fjes_hw_check_mtu(
675 &adapter->hw.ep_shm_info[dest_epid].rx,
677 adapter->stats64.tx_dropped += 1;
678 hw->ep_shm_info[dest_epid].net_stats.tx_dropped += 1;
679 adapter->stats64.tx_errors += 1;
680 hw->ep_shm_info[dest_epid].net_stats.tx_errors += 1;
681 hw->ep_shm_info[dest_epid].ep_stats
682 .tx_dropped_buf_size_mismatch += 1;
686 !fjes_hw_check_vlan_id(
687 &adapter->hw.ep_shm_info[dest_epid].rx,
689 hw->ep_shm_info[dest_epid].ep_stats
690 .tx_dropped_vlanid_mismatch += 1;
693 if (len < VLAN_ETH_HLEN) {
694 memset(shortpkt, 0, VLAN_ETH_HLEN);
695 memcpy(shortpkt, skb->data, skb->len);
700 if (adapter->tx_retry_count == 0) {
701 adapter->tx_start_jiffies = jiffies;
702 adapter->tx_retry_count = 1;
704 adapter->tx_retry_count++;
707 if (fjes_tx_send(adapter, dest_epid, data, len)) {
712 (long)adapter->tx_start_jiffies) >=
713 FJES_TX_RETRY_TIMEOUT) {
714 adapter->stats64.tx_fifo_errors += 1;
715 hw->ep_shm_info[dest_epid].net_stats
716 .tx_fifo_errors += 1;
717 adapter->stats64.tx_errors += 1;
718 hw->ep_shm_info[dest_epid].net_stats
723 netif_trans_update(netdev);
724 hw->ep_shm_info[dest_epid].ep_stats
725 .tx_buffer_full += 1;
726 netif_tx_stop_queue(cur_queue);
728 if (!work_pending(&adapter->tx_stall_task))
729 queue_work(adapter->txrx_wq,
730 &adapter->tx_stall_task);
732 ret = NETDEV_TX_BUSY;
736 adapter->stats64.tx_packets += 1;
737 hw->ep_shm_info[dest_epid].net_stats
739 adapter->stats64.tx_bytes += len;
740 hw->ep_shm_info[dest_epid].net_stats
744 adapter->tx_retry_count = 0;
750 if (ret == NETDEV_TX_OK) {
753 adapter->stats64.tx_packets += 1;
754 hw->ep_shm_info[my_epid].net_stats.tx_packets += 1;
755 adapter->stats64.tx_bytes += 1;
756 hw->ep_shm_info[my_epid].net_stats.tx_bytes += len;
764 fjes_get_stats64(struct net_device *netdev, struct rtnl_link_stats64 *stats)
766 struct fjes_adapter *adapter = netdev_priv(netdev);
768 memcpy(stats, &adapter->stats64, sizeof(struct rtnl_link_stats64));
771 static int fjes_change_mtu(struct net_device *netdev, int new_mtu)
773 struct fjes_adapter *adapter = netdev_priv(netdev);
774 bool running = netif_running(netdev);
775 struct fjes_hw *hw = &adapter->hw;
780 for (idx = 0; fjes_support_mtu[idx] != 0; idx++) {
781 if (new_mtu <= fjes_support_mtu[idx]) {
782 new_mtu = fjes_support_mtu[idx];
783 if (new_mtu == netdev->mtu)
795 spin_lock_irqsave(&hw->rx_status_lock, flags);
796 for (epidx = 0; epidx < hw->max_epid; epidx++) {
797 if (epidx == hw->my_epid)
799 hw->ep_shm_info[epidx].tx.info->v1i.rx_status &=
800 ~FJES_RX_MTU_CHANGING_DONE;
802 spin_unlock_irqrestore(&hw->rx_status_lock, flags);
804 netif_tx_stop_all_queues(netdev);
805 netif_carrier_off(netdev);
806 cancel_work_sync(&adapter->tx_stall_task);
807 napi_disable(&adapter->napi);
811 netif_tx_stop_all_queues(netdev);
814 netdev->mtu = new_mtu;
817 for (epidx = 0; epidx < hw->max_epid; epidx++) {
818 if (epidx == hw->my_epid)
821 spin_lock_irqsave(&hw->rx_status_lock, flags);
822 fjes_hw_setup_epbuf(&hw->ep_shm_info[epidx].tx,
826 hw->ep_shm_info[epidx].tx.info->v1i.rx_status |=
827 FJES_RX_MTU_CHANGING_DONE;
828 spin_unlock_irqrestore(&hw->rx_status_lock, flags);
831 netif_tx_wake_all_queues(netdev);
832 netif_carrier_on(netdev);
833 napi_enable(&adapter->napi);
834 napi_schedule(&adapter->napi);
840 static void fjes_tx_retry(struct net_device *netdev, unsigned int txqueue)
842 struct netdev_queue *queue = netdev_get_tx_queue(netdev, 0);
844 netif_tx_wake_queue(queue);
847 static int fjes_vlan_rx_add_vid(struct net_device *netdev,
848 __be16 proto, u16 vid)
850 struct fjes_adapter *adapter = netdev_priv(netdev);
854 for (epid = 0; epid < adapter->hw.max_epid; epid++) {
855 if (epid == adapter->hw.my_epid)
858 if (!fjes_hw_check_vlan_id(
859 &adapter->hw.ep_shm_info[epid].tx, vid))
860 ret = fjes_hw_set_vlan_id(
861 &adapter->hw.ep_shm_info[epid].tx, vid);
864 return ret ? 0 : -ENOSPC;
867 static int fjes_vlan_rx_kill_vid(struct net_device *netdev,
868 __be16 proto, u16 vid)
870 struct fjes_adapter *adapter = netdev_priv(netdev);
873 for (epid = 0; epid < adapter->hw.max_epid; epid++) {
874 if (epid == adapter->hw.my_epid)
877 fjes_hw_del_vlan_id(&adapter->hw.ep_shm_info[epid].tx, vid);
883 static const struct net_device_ops fjes_netdev_ops = {
884 .ndo_open = fjes_open,
885 .ndo_stop = fjes_close,
886 .ndo_start_xmit = fjes_xmit_frame,
887 .ndo_get_stats64 = fjes_get_stats64,
888 .ndo_change_mtu = fjes_change_mtu,
889 .ndo_tx_timeout = fjes_tx_retry,
890 .ndo_vlan_rx_add_vid = fjes_vlan_rx_add_vid,
891 .ndo_vlan_rx_kill_vid = fjes_vlan_rx_kill_vid,
894 /* fjes_netdev_setup - netdevice initialization routine */
895 static void fjes_netdev_setup(struct net_device *netdev)
899 netdev->watchdog_timeo = FJES_TX_RETRY_INTERVAL;
900 netdev->netdev_ops = &fjes_netdev_ops;
901 fjes_set_ethtool_ops(netdev);
902 netdev->mtu = fjes_support_mtu[3];
903 netdev->min_mtu = fjes_support_mtu[0];
904 netdev->max_mtu = fjes_support_mtu[3];
905 netdev->features |= NETIF_F_HW_VLAN_CTAG_FILTER;
908 static int fjes_rxframe_search_exist(struct fjes_adapter *adapter,
911 struct fjes_hw *hw = &adapter->hw;
912 enum ep_partner_status pstatus;
913 int max_epid, cur_epid;
916 max_epid = hw->max_epid;
917 start_epid = (start_epid + 1 + max_epid) % max_epid;
919 for (i = 0; i < max_epid; i++) {
920 cur_epid = (start_epid + i) % max_epid;
921 if (cur_epid == hw->my_epid)
924 pstatus = fjes_hw_get_partner_ep_status(hw, cur_epid);
925 if (pstatus == EP_PARTNER_SHARED) {
926 if (!fjes_hw_epbuf_rx_is_empty(
927 &hw->ep_shm_info[cur_epid].rx))
934 static void *fjes_rxframe_get(struct fjes_adapter *adapter, size_t *psize,
939 *cur_epid = fjes_rxframe_search_exist(adapter, *cur_epid);
944 fjes_hw_epbuf_rx_curpkt_get_addr(
945 &adapter->hw.ep_shm_info[*cur_epid].rx, psize);
950 static void fjes_rxframe_release(struct fjes_adapter *adapter, int cur_epid)
952 fjes_hw_epbuf_rx_curpkt_drop(&adapter->hw.ep_shm_info[cur_epid].rx);
955 static int fjes_poll(struct napi_struct *napi, int budget)
957 struct fjes_adapter *adapter =
958 container_of(napi, struct fjes_adapter, napi);
959 struct net_device *netdev = napi->dev;
960 struct fjes_hw *hw = &adapter->hw;
968 spin_lock(&hw->rx_status_lock);
969 for (epidx = 0; epidx < hw->max_epid; epidx++) {
970 if (epidx == hw->my_epid)
973 if (fjes_hw_get_partner_ep_status(hw, epidx) ==
975 adapter->hw.ep_shm_info[epidx]
976 .tx.info->v1i.rx_status |= FJES_RX_POLL_WORK;
978 spin_unlock(&hw->rx_status_lock);
980 while (work_done < budget) {
981 prefetch(&adapter->hw);
982 frame = fjes_rxframe_get(adapter, &frame_len, &cur_epid);
985 skb = napi_alloc_skb(napi, frame_len);
987 adapter->stats64.rx_dropped += 1;
988 hw->ep_shm_info[cur_epid].net_stats
990 adapter->stats64.rx_errors += 1;
991 hw->ep_shm_info[cur_epid].net_stats
994 skb_put_data(skb, frame, frame_len);
995 skb->protocol = eth_type_trans(skb, netdev);
996 skb->ip_summed = CHECKSUM_UNNECESSARY;
998 netif_receive_skb(skb);
1002 adapter->stats64.rx_packets += 1;
1003 hw->ep_shm_info[cur_epid].net_stats
1005 adapter->stats64.rx_bytes += frame_len;
1006 hw->ep_shm_info[cur_epid].net_stats
1007 .rx_bytes += frame_len;
1009 if (is_multicast_ether_addr(
1010 ((struct ethhdr *)frame)->h_dest)) {
1011 adapter->stats64.multicast += 1;
1012 hw->ep_shm_info[cur_epid].net_stats
1017 fjes_rxframe_release(adapter, cur_epid);
1018 adapter->unset_rx_last = true;
1024 if (work_done < budget) {
1025 napi_complete_done(napi, work_done);
1027 if (adapter->unset_rx_last) {
1028 adapter->rx_last_jiffies = jiffies;
1029 adapter->unset_rx_last = false;
1032 if (((long)jiffies - (long)adapter->rx_last_jiffies) < 3) {
1033 napi_reschedule(napi);
1035 spin_lock(&hw->rx_status_lock);
1036 for (epidx = 0; epidx < hw->max_epid; epidx++) {
1037 if (epidx == hw->my_epid)
1039 if (fjes_hw_get_partner_ep_status(hw, epidx) ==
1041 adapter->hw.ep_shm_info[epidx].tx
1042 .info->v1i.rx_status &=
1045 spin_unlock(&hw->rx_status_lock);
1047 fjes_hw_set_irqmask(hw, REG_ICTL_MASK_RX_DATA, false);
1054 static int fjes_sw_init(struct fjes_adapter *adapter)
1056 struct net_device *netdev = adapter->netdev;
1058 netif_napi_add(netdev, &adapter->napi, fjes_poll);
1063 static void fjes_force_close_task(struct work_struct *work)
1065 struct fjes_adapter *adapter = container_of(work,
1066 struct fjes_adapter, force_close_task);
1067 struct net_device *netdev = adapter->netdev;
1074 static void fjes_tx_stall_task(struct work_struct *work)
1076 struct fjes_adapter *adapter = container_of(work,
1077 struct fjes_adapter, tx_stall_task);
1078 struct net_device *netdev = adapter->netdev;
1079 struct fjes_hw *hw = &adapter->hw;
1080 int all_queue_available, sendable;
1081 enum ep_partner_status pstatus;
1082 int max_epid, my_epid, epid;
1083 union ep_buffer_info *info;
1086 if (((long)jiffies -
1087 dev_trans_start(netdev)) > FJES_TX_TX_STALL_TIMEOUT) {
1088 netif_wake_queue(netdev);
1092 my_epid = hw->my_epid;
1093 max_epid = hw->max_epid;
1095 for (i = 0; i < 5; i++) {
1096 all_queue_available = 1;
1098 for (epid = 0; epid < max_epid; epid++) {
1099 if (my_epid == epid)
1102 pstatus = fjes_hw_get_partner_ep_status(hw, epid);
1103 sendable = (pstatus == EP_PARTNER_SHARED);
1107 info = adapter->hw.ep_shm_info[epid].tx.info;
1109 if (!(info->v1i.rx_status & FJES_RX_MTU_CHANGING_DONE))
1112 if (EP_RING_FULL(info->v1i.head, info->v1i.tail,
1113 info->v1i.count_max)) {
1114 all_queue_available = 0;
1119 if (all_queue_available) {
1120 netif_wake_queue(netdev);
1125 usleep_range(50, 100);
1127 queue_work(adapter->txrx_wq, &adapter->tx_stall_task);
1130 static void fjes_raise_intr_rxdata_task(struct work_struct *work)
1132 struct fjes_adapter *adapter = container_of(work,
1133 struct fjes_adapter, raise_intr_rxdata_task);
1134 struct fjes_hw *hw = &adapter->hw;
1135 enum ep_partner_status pstatus;
1136 int max_epid, my_epid, epid;
1138 my_epid = hw->my_epid;
1139 max_epid = hw->max_epid;
1141 for (epid = 0; epid < max_epid; epid++)
1142 hw->ep_shm_info[epid].tx_status_work = 0;
1144 for (epid = 0; epid < max_epid; epid++) {
1145 if (epid == my_epid)
1148 pstatus = fjes_hw_get_partner_ep_status(hw, epid);
1149 if (pstatus == EP_PARTNER_SHARED) {
1150 hw->ep_shm_info[epid].tx_status_work =
1151 hw->ep_shm_info[epid].tx.info->v1i.tx_status;
1153 if (hw->ep_shm_info[epid].tx_status_work ==
1154 FJES_TX_DELAY_SEND_PENDING) {
1155 hw->ep_shm_info[epid].tx.info->v1i.tx_status =
1156 FJES_TX_DELAY_SEND_NONE;
1161 for (epid = 0; epid < max_epid; epid++) {
1162 if (epid == my_epid)
1165 pstatus = fjes_hw_get_partner_ep_status(hw, epid);
1166 if ((hw->ep_shm_info[epid].tx_status_work ==
1167 FJES_TX_DELAY_SEND_PENDING) &&
1168 (pstatus == EP_PARTNER_SHARED) &&
1169 !(hw->ep_shm_info[epid].rx.info->v1i.rx_status &
1170 FJES_RX_POLL_WORK)) {
1171 fjes_hw_raise_interrupt(hw, epid,
1172 REG_ICTL_MASK_RX_DATA);
1173 hw->ep_shm_info[epid].ep_stats.send_intr_rx += 1;
1177 usleep_range(500, 1000);
1180 static void fjes_watch_unshare_task(struct work_struct *work)
1182 struct fjes_adapter *adapter =
1183 container_of(work, struct fjes_adapter, unshare_watch_task);
1185 struct net_device *netdev = adapter->netdev;
1186 struct fjes_hw *hw = &adapter->hw;
1188 int unshare_watch, unshare_reserve;
1189 int max_epid, my_epid, epidx;
1190 int stop_req, stop_req_done;
1191 ulong unshare_watch_bitmask;
1192 unsigned long flags;
1197 my_epid = hw->my_epid;
1198 max_epid = hw->max_epid;
1200 unshare_watch_bitmask = adapter->unshare_watch_bitmask;
1201 adapter->unshare_watch_bitmask = 0;
1203 while ((unshare_watch_bitmask || hw->txrx_stop_req_bit) &&
1204 (wait_time < 3000)) {
1205 for (epidx = 0; epidx < max_epid; epidx++) {
1206 if (epidx == my_epid)
1209 is_shared = fjes_hw_epid_is_shared(hw->hw_info.share,
1212 stop_req = test_bit(epidx, &hw->txrx_stop_req_bit);
1214 stop_req_done = hw->ep_shm_info[epidx].rx.info->v1i.rx_status &
1215 FJES_RX_STOP_REQ_DONE;
1217 unshare_watch = test_bit(epidx, &unshare_watch_bitmask);
1219 unshare_reserve = test_bit(epidx,
1220 &hw->hw_info.buffer_unshare_reserve_bit);
1223 (is_shared && (!is_shared || !stop_req_done))) &&
1224 (is_shared || !unshare_watch || !unshare_reserve))
1227 mutex_lock(&hw->hw_info.lock);
1228 ret = fjes_hw_unregister_buff_addr(hw, epidx);
1236 &adapter->force_close_task)) {
1237 adapter->force_reset = true;
1239 &adapter->force_close_task);
1243 mutex_unlock(&hw->hw_info.lock);
1244 hw->ep_shm_info[epidx].ep_stats
1245 .com_unregist_buf_exec += 1;
1247 spin_lock_irqsave(&hw->rx_status_lock, flags);
1248 fjes_hw_setup_epbuf(&hw->ep_shm_info[epidx].tx,
1249 netdev->dev_addr, netdev->mtu);
1250 spin_unlock_irqrestore(&hw->rx_status_lock, flags);
1252 clear_bit(epidx, &hw->txrx_stop_req_bit);
1253 clear_bit(epidx, &unshare_watch_bitmask);
1255 &hw->hw_info.buffer_unshare_reserve_bit);
1262 if (hw->hw_info.buffer_unshare_reserve_bit) {
1263 for (epidx = 0; epidx < max_epid; epidx++) {
1264 if (epidx == my_epid)
1268 &hw->hw_info.buffer_unshare_reserve_bit)) {
1269 mutex_lock(&hw->hw_info.lock);
1271 ret = fjes_hw_unregister_buff_addr(hw, epidx);
1279 &adapter->force_close_task)) {
1280 adapter->force_reset = true;
1282 &adapter->force_close_task);
1286 mutex_unlock(&hw->hw_info.lock);
1288 hw->ep_shm_info[epidx].ep_stats
1289 .com_unregist_buf_exec += 1;
1291 spin_lock_irqsave(&hw->rx_status_lock, flags);
1292 fjes_hw_setup_epbuf(
1293 &hw->ep_shm_info[epidx].tx,
1294 netdev->dev_addr, netdev->mtu);
1295 spin_unlock_irqrestore(&hw->rx_status_lock,
1298 clear_bit(epidx, &hw->txrx_stop_req_bit);
1299 clear_bit(epidx, &unshare_watch_bitmask);
1300 clear_bit(epidx, &hw->hw_info.buffer_unshare_reserve_bit);
1303 if (test_bit(epidx, &unshare_watch_bitmask)) {
1304 spin_lock_irqsave(&hw->rx_status_lock, flags);
1305 hw->ep_shm_info[epidx].tx.info->v1i.rx_status &=
1306 ~FJES_RX_STOP_REQ_DONE;
1307 spin_unlock_irqrestore(&hw->rx_status_lock,
1314 static void fjes_irq_watch_task(struct work_struct *work)
1316 struct fjes_adapter *adapter = container_of(to_delayed_work(work),
1317 struct fjes_adapter, interrupt_watch_task);
1319 local_irq_disable();
1320 fjes_intr(adapter->hw.hw_res.irq, adapter);
1323 if (fjes_rxframe_search_exist(adapter, 0) >= 0)
1324 napi_schedule(&adapter->napi);
1326 if (adapter->interrupt_watch_enable) {
1327 if (!delayed_work_pending(&adapter->interrupt_watch_task))
1328 queue_delayed_work(adapter->control_wq,
1329 &adapter->interrupt_watch_task,
1330 FJES_IRQ_WATCH_DELAY);
1334 /* fjes_probe - Device Initialization Routine */
1335 static int fjes_probe(struct platform_device *plat_dev)
1337 struct fjes_adapter *adapter;
1338 struct net_device *netdev;
1339 struct resource *res;
1345 netdev = alloc_netdev_mq(sizeof(struct fjes_adapter), "es%d",
1346 NET_NAME_UNKNOWN, fjes_netdev_setup,
1352 SET_NETDEV_DEV(netdev, &plat_dev->dev);
1354 dev_set_drvdata(&plat_dev->dev, netdev);
1355 adapter = netdev_priv(netdev);
1356 adapter->netdev = netdev;
1357 adapter->plat_dev = plat_dev;
1361 /* setup the private structure */
1362 err = fjes_sw_init(adapter);
1364 goto err_free_netdev;
1366 INIT_WORK(&adapter->force_close_task, fjes_force_close_task);
1367 adapter->force_reset = false;
1368 adapter->open_guard = false;
1370 adapter->txrx_wq = alloc_workqueue(DRV_NAME "/txrx", WQ_MEM_RECLAIM, 0);
1371 if (unlikely(!adapter->txrx_wq)) {
1373 goto err_free_netdev;
1376 adapter->control_wq = alloc_workqueue(DRV_NAME "/control",
1378 if (unlikely(!adapter->control_wq)) {
1380 goto err_free_txrx_wq;
1383 INIT_WORK(&adapter->tx_stall_task, fjes_tx_stall_task);
1384 INIT_WORK(&adapter->raise_intr_rxdata_task,
1385 fjes_raise_intr_rxdata_task);
1386 INIT_WORK(&adapter->unshare_watch_task, fjes_watch_unshare_task);
1387 adapter->unshare_watch_bitmask = 0;
1389 INIT_DELAYED_WORK(&adapter->interrupt_watch_task, fjes_irq_watch_task);
1390 adapter->interrupt_watch_enable = false;
1392 res = platform_get_resource(plat_dev, IORESOURCE_MEM, 0);
1395 goto err_free_control_wq;
1397 hw->hw_res.start = res->start;
1398 hw->hw_res.size = resource_size(res);
1399 hw->hw_res.irq = platform_get_irq(plat_dev, 0);
1400 if (hw->hw_res.irq < 0) {
1401 err = hw->hw_res.irq;
1402 goto err_free_control_wq;
1405 err = fjes_hw_init(&adapter->hw);
1407 goto err_free_control_wq;
1409 /* setup MAC address (02:00:00:00:00:[epid])*/
1415 addr[5] = hw->my_epid; /* EPID */
1416 eth_hw_addr_set(netdev, addr);
1418 err = register_netdev(netdev);
1422 netif_carrier_off(netdev);
1424 fjes_dbg_adapter_init(adapter);
1429 fjes_hw_exit(&adapter->hw);
1430 err_free_control_wq:
1431 destroy_workqueue(adapter->control_wq);
1433 destroy_workqueue(adapter->txrx_wq);
1435 free_netdev(netdev);
1440 /* fjes_remove - Device Removal Routine */
1441 static int fjes_remove(struct platform_device *plat_dev)
1443 struct net_device *netdev = dev_get_drvdata(&plat_dev->dev);
1444 struct fjes_adapter *adapter = netdev_priv(netdev);
1445 struct fjes_hw *hw = &adapter->hw;
1447 fjes_dbg_adapter_exit(adapter);
1449 cancel_delayed_work_sync(&adapter->interrupt_watch_task);
1450 cancel_work_sync(&adapter->unshare_watch_task);
1451 cancel_work_sync(&adapter->raise_intr_rxdata_task);
1452 cancel_work_sync(&adapter->tx_stall_task);
1453 if (adapter->control_wq)
1454 destroy_workqueue(adapter->control_wq);
1455 if (adapter->txrx_wq)
1456 destroy_workqueue(adapter->txrx_wq);
1458 unregister_netdev(netdev);
1462 netif_napi_del(&adapter->napi);
1464 free_netdev(netdev);
1469 static struct platform_driver fjes_driver = {
1473 .probe = fjes_probe,
1474 .remove = fjes_remove,
1478 acpi_find_extended_socket_device(acpi_handle obj_handle, u32 level,
1479 void *context, void **return_value)
1481 struct acpi_device *device;
1482 bool *found = context;
1484 device = acpi_fetch_acpi_dev(obj_handle);
1488 if (strcmp(acpi_device_hid(device), ACPI_MOTHERBOARD_RESOURCE_HID))
1491 if (!is_extended_socket_device(device))
1494 if (acpi_check_extended_socket_status(device))
1498 return AE_CTRL_TERMINATE;
1501 /* fjes_init_module - Driver Registration Routine */
1502 static int __init fjes_init_module(void)
1507 acpi_walk_namespace(ACPI_TYPE_DEVICE, ACPI_ROOT_OBJECT, ACPI_UINT32_MAX,
1508 acpi_find_extended_socket_device, NULL, &found,
1514 pr_info("%s - version %s - %s\n",
1515 fjes_driver_string, fjes_driver_version, fjes_copyright);
1519 result = platform_driver_register(&fjes_driver);
1525 result = acpi_bus_register_driver(&fjes_acpi_driver);
1527 goto fail_acpi_driver;
1532 platform_driver_unregister(&fjes_driver);
1537 module_init(fjes_init_module);
1539 /* fjes_exit_module - Driver Exit Cleanup Routine */
1540 static void __exit fjes_exit_module(void)
1542 acpi_bus_unregister_driver(&fjes_acpi_driver);
1543 platform_driver_unregister(&fjes_driver);
1547 module_exit(fjes_exit_module);