xgene_enet_xgmac.c 15 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574
  1. /* Applied Micro X-Gene SoC Ethernet Driver
  2. *
  3. * Copyright (c) 2014, Applied Micro Circuits Corporation
  4. * Authors: Iyappan Subramanian <isubramanian@apm.com>
  5. * Keyur Chudgar <kchudgar@apm.com>
  6. *
  7. * This program is free software; you can redistribute it and/or modify it
  8. * under the terms of the GNU General Public License as published by the
  9. * Free Software Foundation; either version 2 of the License, or (at your
  10. * option) any later version.
  11. *
  12. * This program is distributed in the hope that it will be useful,
  13. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  14. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
  15. * GNU General Public License for more details.
  16. *
  17. * You should have received a copy of the GNU General Public License
  18. * along with this program. If not, see <http://www.gnu.org/licenses/>.
  19. */
  20. #include <linux/of_gpio.h>
  21. #include <linux/gpio.h>
  22. #include "xgene_enet_main.h"
  23. #include "xgene_enet_hw.h"
  24. #include "xgene_enet_xgmac.h"
  25. static void xgene_enet_wr_csr(struct xgene_enet_pdata *pdata,
  26. u32 offset, u32 val)
  27. {
  28. void __iomem *addr = pdata->eth_csr_addr + offset;
  29. iowrite32(val, addr);
  30. }
  31. static void xgene_enet_wr_ring_if(struct xgene_enet_pdata *pdata,
  32. u32 offset, u32 val)
  33. {
  34. void __iomem *addr = pdata->eth_ring_if_addr + offset;
  35. iowrite32(val, addr);
  36. }
  37. static void xgene_enet_wr_diag_csr(struct xgene_enet_pdata *pdata,
  38. u32 offset, u32 val)
  39. {
  40. void __iomem *addr = pdata->eth_diag_csr_addr + offset;
  41. iowrite32(val, addr);
  42. }
  43. static bool xgene_enet_wr_indirect(void __iomem *addr, void __iomem *wr,
  44. void __iomem *cmd, void __iomem *cmd_done,
  45. u32 wr_addr, u32 wr_data)
  46. {
  47. u32 done;
  48. u8 wait = 10;
  49. iowrite32(wr_addr, addr);
  50. iowrite32(wr_data, wr);
  51. iowrite32(XGENE_ENET_WR_CMD, cmd);
  52. /* wait for write command to complete */
  53. while (!(done = ioread32(cmd_done)) && wait--)
  54. udelay(1);
  55. if (!done)
  56. return false;
  57. iowrite32(0, cmd);
  58. return true;
  59. }
  60. static void xgene_enet_wr_mac(struct xgene_enet_pdata *pdata,
  61. u32 wr_addr, u32 wr_data)
  62. {
  63. void __iomem *addr, *wr, *cmd, *cmd_done;
  64. addr = pdata->mcx_mac_addr + MAC_ADDR_REG_OFFSET;
  65. wr = pdata->mcx_mac_addr + MAC_WRITE_REG_OFFSET;
  66. cmd = pdata->mcx_mac_addr + MAC_COMMAND_REG_OFFSET;
  67. cmd_done = pdata->mcx_mac_addr + MAC_COMMAND_DONE_REG_OFFSET;
  68. if (!xgene_enet_wr_indirect(addr, wr, cmd, cmd_done, wr_addr, wr_data))
  69. netdev_err(pdata->ndev, "MCX mac write failed, addr: %04x\n",
  70. wr_addr);
  71. }
  72. static void xgene_enet_wr_pcs(struct xgene_enet_pdata *pdata,
  73. u32 wr_addr, u32 wr_data)
  74. {
  75. void __iomem *addr, *wr, *cmd, *cmd_done;
  76. addr = pdata->pcs_addr + PCS_ADDR_REG_OFFSET;
  77. wr = pdata->pcs_addr + PCS_WRITE_REG_OFFSET;
  78. cmd = pdata->pcs_addr + PCS_COMMAND_REG_OFFSET;
  79. cmd_done = pdata->pcs_addr + PCS_COMMAND_DONE_REG_OFFSET;
  80. if (!xgene_enet_wr_indirect(addr, wr, cmd, cmd_done, wr_addr, wr_data))
  81. netdev_err(pdata->ndev, "PCS write failed, addr: %04x\n",
  82. wr_addr);
  83. }
  84. static void xgene_enet_wr_axg_csr(struct xgene_enet_pdata *pdata,
  85. u32 offset, u32 val)
  86. {
  87. void __iomem *addr = pdata->mcx_mac_csr_addr + offset;
  88. iowrite32(val, addr);
  89. }
  90. static void xgene_enet_rd_csr(struct xgene_enet_pdata *pdata,
  91. u32 offset, u32 *val)
  92. {
  93. void __iomem *addr = pdata->eth_csr_addr + offset;
  94. *val = ioread32(addr);
  95. }
  96. static void xgene_enet_rd_diag_csr(struct xgene_enet_pdata *pdata,
  97. u32 offset, u32 *val)
  98. {
  99. void __iomem *addr = pdata->eth_diag_csr_addr + offset;
  100. *val = ioread32(addr);
  101. }
  102. static bool xgene_enet_rd_indirect(void __iomem *addr, void __iomem *rd,
  103. void __iomem *cmd, void __iomem *cmd_done,
  104. u32 rd_addr, u32 *rd_data)
  105. {
  106. u32 done;
  107. u8 wait = 10;
  108. iowrite32(rd_addr, addr);
  109. iowrite32(XGENE_ENET_RD_CMD, cmd);
  110. /* wait for read command to complete */
  111. while (!(done = ioread32(cmd_done)) && wait--)
  112. udelay(1);
  113. if (!done)
  114. return false;
  115. *rd_data = ioread32(rd);
  116. iowrite32(0, cmd);
  117. return true;
  118. }
  119. static void xgene_enet_rd_mac(struct xgene_enet_pdata *pdata,
  120. u32 rd_addr, u32 *rd_data)
  121. {
  122. void __iomem *addr, *rd, *cmd, *cmd_done;
  123. addr = pdata->mcx_mac_addr + MAC_ADDR_REG_OFFSET;
  124. rd = pdata->mcx_mac_addr + MAC_READ_REG_OFFSET;
  125. cmd = pdata->mcx_mac_addr + MAC_COMMAND_REG_OFFSET;
  126. cmd_done = pdata->mcx_mac_addr + MAC_COMMAND_DONE_REG_OFFSET;
  127. if (!xgene_enet_rd_indirect(addr, rd, cmd, cmd_done, rd_addr, rd_data))
  128. netdev_err(pdata->ndev, "MCX mac read failed, addr: %04x\n",
  129. rd_addr);
  130. }
  131. static bool xgene_enet_rd_pcs(struct xgene_enet_pdata *pdata,
  132. u32 rd_addr, u32 *rd_data)
  133. {
  134. void __iomem *addr, *rd, *cmd, *cmd_done;
  135. bool success;
  136. addr = pdata->pcs_addr + PCS_ADDR_REG_OFFSET;
  137. rd = pdata->pcs_addr + PCS_READ_REG_OFFSET;
  138. cmd = pdata->pcs_addr + PCS_COMMAND_REG_OFFSET;
  139. cmd_done = pdata->pcs_addr + PCS_COMMAND_DONE_REG_OFFSET;
  140. success = xgene_enet_rd_indirect(addr, rd, cmd, cmd_done, rd_addr, rd_data);
  141. if (!success)
  142. netdev_err(pdata->ndev, "PCS read failed, addr: %04x\n",
  143. rd_addr);
  144. return success;
  145. }
  146. static void xgene_enet_rd_axg_csr(struct xgene_enet_pdata *pdata,
  147. u32 offset, u32 *val)
  148. {
  149. void __iomem *addr = pdata->mcx_mac_csr_addr + offset;
  150. *val = ioread32(addr);
  151. }
  152. static int xgene_enet_ecc_init(struct xgene_enet_pdata *pdata)
  153. {
  154. struct net_device *ndev = pdata->ndev;
  155. u32 data;
  156. u8 wait = 10;
  157. xgene_enet_wr_diag_csr(pdata, ENET_CFG_MEM_RAM_SHUTDOWN_ADDR, 0x0);
  158. do {
  159. usleep_range(100, 110);
  160. xgene_enet_rd_diag_csr(pdata, ENET_BLOCK_MEM_RDY_ADDR, &data);
  161. } while ((data != 0xffffffff) && wait--);
  162. if (data != 0xffffffff) {
  163. netdev_err(ndev, "Failed to release memory from shutdown\n");
  164. return -ENODEV;
  165. }
  166. return 0;
  167. }
  168. static void xgene_enet_config_ring_if_assoc(struct xgene_enet_pdata *pdata)
  169. {
  170. xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIWQASSOC_ADDR, 0);
  171. xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIFPQASSOC_ADDR, 0);
  172. xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIQMLITEWQASSOC_ADDR, 0);
  173. xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIQMLITEFPQASSOC_ADDR, 0);
  174. }
  175. static void xgene_xgmac_reset(struct xgene_enet_pdata *pdata)
  176. {
  177. xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_0, HSTMACRST);
  178. xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_0, 0);
  179. }
  180. static void xgene_pcs_reset(struct xgene_enet_pdata *pdata)
  181. {
  182. u32 data;
  183. if (!xgene_enet_rd_pcs(pdata, PCS_CONTROL_1, &data))
  184. return;
  185. xgene_enet_wr_pcs(pdata, PCS_CONTROL_1, data | PCS_CTRL_PCS_RST);
  186. xgene_enet_wr_pcs(pdata, PCS_CONTROL_1, data & ~PCS_CTRL_PCS_RST);
  187. }
  188. static void xgene_xgmac_set_mac_addr(struct xgene_enet_pdata *pdata)
  189. {
  190. u32 addr0, addr1;
  191. u8 *dev_addr = pdata->ndev->dev_addr;
  192. addr0 = (dev_addr[3] << 24) | (dev_addr[2] << 16) |
  193. (dev_addr[1] << 8) | dev_addr[0];
  194. addr1 = (dev_addr[5] << 24) | (dev_addr[4] << 16);
  195. xgene_enet_wr_mac(pdata, HSTMACADR_LSW_ADDR, addr0);
  196. xgene_enet_wr_mac(pdata, HSTMACADR_MSW_ADDR, addr1);
  197. }
  198. static void xgene_xgmac_set_mss(struct xgene_enet_pdata *pdata,
  199. u16 mss, u8 index)
  200. {
  201. u8 offset;
  202. u32 data;
  203. offset = (index < 2) ? 0 : 4;
  204. xgene_enet_rd_csr(pdata, XG_TSIF_MSS_REG0_ADDR + offset, &data);
  205. if (!(index & 0x1))
  206. data = SET_VAL(TSO_MSS1, data >> TSO_MSS1_POS) |
  207. SET_VAL(TSO_MSS0, mss);
  208. else
  209. data = SET_VAL(TSO_MSS1, mss) | SET_VAL(TSO_MSS0, data);
  210. xgene_enet_wr_csr(pdata, XG_TSIF_MSS_REG0_ADDR + offset, data);
  211. }
  212. static void xgene_xgmac_set_frame_size(struct xgene_enet_pdata *pdata, int size)
  213. {
  214. xgene_enet_wr_mac(pdata, HSTMAXFRAME_LENGTH_ADDR,
  215. ((((size + 2) >> 2) << 16) | size));
  216. }
  217. static u32 xgene_enet_link_status(struct xgene_enet_pdata *pdata)
  218. {
  219. u32 data;
  220. xgene_enet_rd_csr(pdata, XG_LINK_STATUS_ADDR, &data);
  221. return data;
  222. }
  223. static void xgene_xgmac_enable_tx_pause(struct xgene_enet_pdata *pdata,
  224. bool enable)
  225. {
  226. u32 data;
  227. xgene_enet_rd_axg_csr(pdata, XGENET_CSR_ECM_CFG_0_ADDR, &data);
  228. if (enable)
  229. data |= MULTI_DPF_AUTOCTRL | PAUSE_XON_EN;
  230. else
  231. data &= ~(MULTI_DPF_AUTOCTRL | PAUSE_XON_EN);
  232. xgene_enet_wr_axg_csr(pdata, XGENET_CSR_ECM_CFG_0_ADDR, data);
  233. }
  234. static void xgene_xgmac_flowctl_tx(struct xgene_enet_pdata *pdata, bool enable)
  235. {
  236. u32 data;
  237. xgene_enet_rd_mac(pdata, AXGMAC_CONFIG_1, &data);
  238. if (enable)
  239. data |= HSTTCTLEN;
  240. else
  241. data &= ~HSTTCTLEN;
  242. xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_1, data);
  243. pdata->mac_ops->enable_tx_pause(pdata, enable);
  244. }
  245. static void xgene_xgmac_flowctl_rx(struct xgene_enet_pdata *pdata, bool enable)
  246. {
  247. u32 data;
  248. xgene_enet_rd_mac(pdata, AXGMAC_CONFIG_1, &data);
  249. if (enable)
  250. data |= HSTRCTLEN;
  251. else
  252. data &= ~HSTRCTLEN;
  253. xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_1, data);
  254. }
  255. static void xgene_xgmac_init(struct xgene_enet_pdata *pdata)
  256. {
  257. u32 data;
  258. xgene_xgmac_reset(pdata);
  259. xgene_enet_rd_mac(pdata, AXGMAC_CONFIG_1, &data);
  260. data |= HSTPPEN;
  261. data &= ~HSTLENCHK;
  262. xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_1, data);
  263. xgene_xgmac_set_mac_addr(pdata);
  264. xgene_enet_rd_csr(pdata, XG_RSIF_CONFIG_REG_ADDR, &data);
  265. data |= CFG_RSIF_FPBUFF_TIMEOUT_EN;
  266. xgene_enet_wr_csr(pdata, XG_RSIF_CONFIG_REG_ADDR, data);
  267. xgene_enet_rd_csr(pdata, XG_ENET_SPARE_CFG_REG_ADDR, &data);
  268. data |= BIT(12);
  269. xgene_enet_wr_csr(pdata, XG_ENET_SPARE_CFG_REG_ADDR, data);
  270. xgene_enet_wr_csr(pdata, XG_ENET_SPARE_CFG_REG_1_ADDR, 0x82);
  271. xgene_enet_wr_csr(pdata, XGENET_RX_DV_GATE_REG_0_ADDR, 0);
  272. xgene_enet_wr_csr(pdata, XG_CFG_BYPASS_ADDR, RESUME_TX);
  273. /* Configure HW pause frame generation */
  274. xgene_enet_rd_axg_csr(pdata, XGENET_CSR_MULTI_DPF0_ADDR, &data);
  275. data = (DEF_QUANTA << 16) | (data & 0xFFFF);
  276. xgene_enet_wr_axg_csr(pdata, XGENET_CSR_MULTI_DPF0_ADDR, data);
  277. if (pdata->enet_id != XGENE_ENET1) {
  278. xgene_enet_rd_axg_csr(pdata, XGENET_CSR_MULTI_DPF1_ADDR, &data);
  279. data = (NORM_PAUSE_OPCODE << 16) | (data & 0xFFFF);
  280. xgene_enet_wr_axg_csr(pdata, XGENET_CSR_MULTI_DPF1_ADDR, data);
  281. }
  282. data = (XG_DEF_PAUSE_OFF_THRES << 16) | XG_DEF_PAUSE_THRES;
  283. xgene_enet_wr_csr(pdata, XG_RXBUF_PAUSE_THRESH, data);
  284. xgene_xgmac_flowctl_tx(pdata, pdata->tx_pause);
  285. xgene_xgmac_flowctl_rx(pdata, pdata->rx_pause);
  286. }
  287. static void xgene_xgmac_rx_enable(struct xgene_enet_pdata *pdata)
  288. {
  289. u32 data;
  290. xgene_enet_rd_mac(pdata, AXGMAC_CONFIG_1, &data);
  291. xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_1, data | HSTRFEN);
  292. }
  293. static void xgene_xgmac_tx_enable(struct xgene_enet_pdata *pdata)
  294. {
  295. u32 data;
  296. xgene_enet_rd_mac(pdata, AXGMAC_CONFIG_1, &data);
  297. xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_1, data | HSTTFEN);
  298. }
  299. static void xgene_xgmac_rx_disable(struct xgene_enet_pdata *pdata)
  300. {
  301. u32 data;
  302. xgene_enet_rd_mac(pdata, AXGMAC_CONFIG_1, &data);
  303. xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_1, data & ~HSTRFEN);
  304. }
  305. static void xgene_xgmac_tx_disable(struct xgene_enet_pdata *pdata)
  306. {
  307. u32 data;
  308. xgene_enet_rd_mac(pdata, AXGMAC_CONFIG_1, &data);
  309. xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_1, data & ~HSTTFEN);
  310. }
  311. static int xgene_enet_reset(struct xgene_enet_pdata *pdata)
  312. {
  313. struct device *dev = &pdata->pdev->dev;
  314. if (!xgene_ring_mgr_init(pdata))
  315. return -ENODEV;
  316. if (dev->of_node) {
  317. clk_prepare_enable(pdata->clk);
  318. udelay(5);
  319. clk_disable_unprepare(pdata->clk);
  320. udelay(5);
  321. clk_prepare_enable(pdata->clk);
  322. udelay(5);
  323. } else {
  324. #ifdef CONFIG_ACPI
  325. if (acpi_has_method(ACPI_HANDLE(&pdata->pdev->dev), "_RST")) {
  326. acpi_evaluate_object(ACPI_HANDLE(&pdata->pdev->dev),
  327. "_RST", NULL, NULL);
  328. } else if (acpi_has_method(ACPI_HANDLE(&pdata->pdev->dev),
  329. "_INI")) {
  330. acpi_evaluate_object(ACPI_HANDLE(&pdata->pdev->dev),
  331. "_INI", NULL, NULL);
  332. }
  333. #endif
  334. }
  335. xgene_enet_ecc_init(pdata);
  336. xgene_enet_config_ring_if_assoc(pdata);
  337. return 0;
  338. }
  339. static void xgene_enet_xgcle_bypass(struct xgene_enet_pdata *pdata,
  340. u32 dst_ring_num, u16 bufpool_id,
  341. u16 nxtbufpool_id)
  342. {
  343. u32 cb, fpsel, nxtfpsel;
  344. xgene_enet_rd_csr(pdata, XCLE_BYPASS_REG0_ADDR, &cb);
  345. cb |= CFG_CLE_BYPASS_EN0;
  346. CFG_CLE_IP_PROTOCOL0_SET(&cb, 3);
  347. xgene_enet_wr_csr(pdata, XCLE_BYPASS_REG0_ADDR, cb);
  348. fpsel = xgene_enet_get_fpsel(bufpool_id);
  349. nxtfpsel = xgene_enet_get_fpsel(nxtbufpool_id);
  350. xgene_enet_rd_csr(pdata, XCLE_BYPASS_REG1_ADDR, &cb);
  351. CFG_CLE_DSTQID0_SET(&cb, dst_ring_num);
  352. CFG_CLE_FPSEL0_SET(&cb, fpsel);
  353. CFG_CLE_NXTFPSEL0_SET(&cb, nxtfpsel);
  354. xgene_enet_wr_csr(pdata, XCLE_BYPASS_REG1_ADDR, cb);
  355. pr_info("+ cle_bypass: fpsel: %d nxtfpsel: %d\n", fpsel, nxtfpsel);
  356. }
  357. static void xgene_enet_shutdown(struct xgene_enet_pdata *pdata)
  358. {
  359. struct device *dev = &pdata->pdev->dev;
  360. struct xgene_enet_desc_ring *ring;
  361. u32 pb;
  362. int i;
  363. pb = 0;
  364. for (i = 0; i < pdata->rxq_cnt; i++) {
  365. ring = pdata->rx_ring[i]->buf_pool;
  366. pb |= BIT(xgene_enet_get_fpsel(ring->id));
  367. ring = pdata->rx_ring[i]->page_pool;
  368. if (ring)
  369. pb |= BIT(xgene_enet_get_fpsel(ring->id));
  370. }
  371. xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIFPRESET_ADDR, pb);
  372. pb = 0;
  373. for (i = 0; i < pdata->txq_cnt; i++) {
  374. ring = pdata->tx_ring[i];
  375. pb |= BIT(xgene_enet_ring_bufnum(ring->id));
  376. }
  377. xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIWQRESET_ADDR, pb);
  378. if (dev->of_node) {
  379. if (!IS_ERR(pdata->clk))
  380. clk_disable_unprepare(pdata->clk);
  381. }
  382. }
  383. static void xgene_enet_clear(struct xgene_enet_pdata *pdata,
  384. struct xgene_enet_desc_ring *ring)
  385. {
  386. u32 addr, data;
  387. if (xgene_enet_is_bufpool(ring->id)) {
  388. addr = ENET_CFGSSQMIFPRESET_ADDR;
  389. data = BIT(xgene_enet_get_fpsel(ring->id));
  390. } else {
  391. addr = ENET_CFGSSQMIWQRESET_ADDR;
  392. data = BIT(xgene_enet_ring_bufnum(ring->id));
  393. }
  394. xgene_enet_wr_ring_if(pdata, addr, data);
  395. }
  396. static int xgene_enet_gpio_lookup(struct xgene_enet_pdata *pdata)
  397. {
  398. struct device *dev = &pdata->pdev->dev;
  399. pdata->sfp_rdy = gpiod_get(dev, "rxlos", GPIOD_IN);
  400. if (IS_ERR(pdata->sfp_rdy))
  401. pdata->sfp_rdy = gpiod_get(dev, "sfp", GPIOD_IN);
  402. if (IS_ERR(pdata->sfp_rdy))
  403. return -ENODEV;
  404. return 0;
  405. }
  406. static void xgene_enet_link_state(struct work_struct *work)
  407. {
  408. struct xgene_enet_pdata *pdata = container_of(to_delayed_work(work),
  409. struct xgene_enet_pdata, link_work);
  410. struct net_device *ndev = pdata->ndev;
  411. u32 link_status, poll_interval;
  412. link_status = xgene_enet_link_status(pdata);
  413. if (pdata->sfp_gpio_en && link_status &&
  414. (!IS_ERR(pdata->sfp_rdy) || !xgene_enet_gpio_lookup(pdata)) &&
  415. !gpiod_get_value(pdata->sfp_rdy))
  416. link_status = 0;
  417. if (link_status) {
  418. if (!netif_carrier_ok(ndev)) {
  419. netif_carrier_on(ndev);
  420. xgene_xgmac_rx_enable(pdata);
  421. xgene_xgmac_tx_enable(pdata);
  422. netdev_info(ndev, "Link is Up - 10Gbps\n");
  423. }
  424. poll_interval = PHY_POLL_LINK_ON;
  425. } else {
  426. if (netif_carrier_ok(ndev)) {
  427. xgene_xgmac_rx_disable(pdata);
  428. xgene_xgmac_tx_disable(pdata);
  429. netif_carrier_off(ndev);
  430. netdev_info(ndev, "Link is Down\n");
  431. }
  432. poll_interval = PHY_POLL_LINK_OFF;
  433. xgene_pcs_reset(pdata);
  434. }
  435. schedule_delayed_work(&pdata->link_work, poll_interval);
  436. }
  437. const struct xgene_mac_ops xgene_xgmac_ops = {
  438. .init = xgene_xgmac_init,
  439. .reset = xgene_xgmac_reset,
  440. .rx_enable = xgene_xgmac_rx_enable,
  441. .tx_enable = xgene_xgmac_tx_enable,
  442. .rx_disable = xgene_xgmac_rx_disable,
  443. .tx_disable = xgene_xgmac_tx_disable,
  444. .set_mac_addr = xgene_xgmac_set_mac_addr,
  445. .set_framesize = xgene_xgmac_set_frame_size,
  446. .set_mss = xgene_xgmac_set_mss,
  447. .link_state = xgene_enet_link_state,
  448. .enable_tx_pause = xgene_xgmac_enable_tx_pause,
  449. .flowctl_rx = xgene_xgmac_flowctl_rx,
  450. .flowctl_tx = xgene_xgmac_flowctl_tx
  451. };
  452. const struct xgene_port_ops xgene_xgport_ops = {
  453. .reset = xgene_enet_reset,
  454. .clear = xgene_enet_clear,
  455. .cle_bypass = xgene_enet_xgcle_bypass,
  456. .shutdown = xgene_enet_shutdown,
  457. };