qlge_dbg.c 61 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625162616271628162916301631163216331634163516361637163816391640164116421643164416451646164716481649165016511652165316541655165616571658165916601661166216631664166516661667166816691670167116721673167416751676167716781679168016811682168316841685168616871688168916901691169216931694169516961697169816991700170117021703170417051706170717081709171017111712171317141715171617171718171917201721172217231724172517261727172817291730173117321733173417351736173717381739174017411742174317441745174617471748174917501751175217531754175517561757175817591760176117621763176417651766176717681769177017711772177317741775177617771778177917801781178217831784178517861787178817891790179117921793179417951796179717981799180018011802180318041805180618071808180918101811181218131814181518161817181818191820182118221823182418251826182718281829183018311832183318341835183618371838183918401841184218431844184518461847184818491850185118521853185418551856185718581859186018611862186318641865186618671868186918701871187218731874187518761877187818791880188118821883188418851886188718881889189018911892189318941895189618971898189919001901190219031904190519061907190819091910191119121913191419151916191719181919192019211922192319241925192619271928192919301931193219331934193519361937193819391940194119421943194419451946194719481949195019511952195319541955195619571958195919601961196219631964196519661967196819691970197119721973197419751976197719781979198019811982198319841985198619871988198919901991199219931994199519961997199819992000200120022003200420052006200720082009201020112012201320142015201620172018201920202021202220232024
  1. // SPDX-License-Identifier: GPL-2.0
  2. #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
  3. #include <linux/slab.h>
  4. #include "qlge.h"
  5. /* Read a NIC register from the alternate function. */
  6. static u32 ql_read_other_func_reg(struct ql_adapter *qdev,
  7. u32 reg)
  8. {
  9. u32 register_to_read;
  10. u32 reg_val;
  11. unsigned int status = 0;
  12. register_to_read = MPI_NIC_REG_BLOCK
  13. | MPI_NIC_READ
  14. | (qdev->alt_func << MPI_NIC_FUNCTION_SHIFT)
  15. | reg;
  16. status = ql_read_mpi_reg(qdev, register_to_read, &reg_val);
  17. if (status != 0)
  18. return 0xffffffff;
  19. return reg_val;
  20. }
  21. /* Write a NIC register from the alternate function. */
  22. static int ql_write_other_func_reg(struct ql_adapter *qdev,
  23. u32 reg, u32 reg_val)
  24. {
  25. u32 register_to_read;
  26. int status = 0;
  27. register_to_read = MPI_NIC_REG_BLOCK
  28. | MPI_NIC_READ
  29. | (qdev->alt_func << MPI_NIC_FUNCTION_SHIFT)
  30. | reg;
  31. status = ql_write_mpi_reg(qdev, register_to_read, reg_val);
  32. return status;
  33. }
  34. static int ql_wait_other_func_reg_rdy(struct ql_adapter *qdev, u32 reg,
  35. u32 bit, u32 err_bit)
  36. {
  37. u32 temp;
  38. int count = 10;
  39. while (count) {
  40. temp = ql_read_other_func_reg(qdev, reg);
  41. /* check for errors */
  42. if (temp & err_bit)
  43. return -1;
  44. else if (temp & bit)
  45. return 0;
  46. mdelay(10);
  47. count--;
  48. }
  49. return -1;
  50. }
  51. static int ql_read_other_func_serdes_reg(struct ql_adapter *qdev, u32 reg,
  52. u32 *data)
  53. {
  54. int status;
  55. /* wait for reg to come ready */
  56. status = ql_wait_other_func_reg_rdy(qdev, XG_SERDES_ADDR / 4,
  57. XG_SERDES_ADDR_RDY, 0);
  58. if (status)
  59. goto exit;
  60. /* set up for reg read */
  61. ql_write_other_func_reg(qdev, XG_SERDES_ADDR/4, reg | PROC_ADDR_R);
  62. /* wait for reg to come ready */
  63. status = ql_wait_other_func_reg_rdy(qdev, XG_SERDES_ADDR / 4,
  64. XG_SERDES_ADDR_RDY, 0);
  65. if (status)
  66. goto exit;
  67. /* get the data */
  68. *data = ql_read_other_func_reg(qdev, (XG_SERDES_DATA / 4));
  69. exit:
  70. return status;
  71. }
  72. /* Read out the SERDES registers */
  73. static int ql_read_serdes_reg(struct ql_adapter *qdev, u32 reg, u32 *data)
  74. {
  75. int status;
  76. /* wait for reg to come ready */
  77. status = ql_wait_reg_rdy(qdev, XG_SERDES_ADDR, XG_SERDES_ADDR_RDY, 0);
  78. if (status)
  79. goto exit;
  80. /* set up for reg read */
  81. ql_write32(qdev, XG_SERDES_ADDR, reg | PROC_ADDR_R);
  82. /* wait for reg to come ready */
  83. status = ql_wait_reg_rdy(qdev, XG_SERDES_ADDR, XG_SERDES_ADDR_RDY, 0);
  84. if (status)
  85. goto exit;
  86. /* get the data */
  87. *data = ql_read32(qdev, XG_SERDES_DATA);
  88. exit:
  89. return status;
  90. }
  91. static void ql_get_both_serdes(struct ql_adapter *qdev, u32 addr,
  92. u32 *direct_ptr, u32 *indirect_ptr,
  93. unsigned int direct_valid, unsigned int indirect_valid)
  94. {
  95. unsigned int status;
  96. status = 1;
  97. if (direct_valid)
  98. status = ql_read_serdes_reg(qdev, addr, direct_ptr);
  99. /* Dead fill any failures or invalids. */
  100. if (status)
  101. *direct_ptr = 0xDEADBEEF;
  102. status = 1;
  103. if (indirect_valid)
  104. status = ql_read_other_func_serdes_reg(
  105. qdev, addr, indirect_ptr);
  106. /* Dead fill any failures or invalids. */
  107. if (status)
  108. *indirect_ptr = 0xDEADBEEF;
  109. }
  110. static int ql_get_serdes_regs(struct ql_adapter *qdev,
  111. struct ql_mpi_coredump *mpi_coredump)
  112. {
  113. int status;
  114. unsigned int xfi_direct_valid, xfi_indirect_valid, xaui_direct_valid;
  115. unsigned int xaui_indirect_valid, i;
  116. u32 *direct_ptr, temp;
  117. u32 *indirect_ptr;
  118. xfi_direct_valid = xfi_indirect_valid = 0;
  119. xaui_direct_valid = xaui_indirect_valid = 1;
  120. /* The XAUI needs to be read out per port */
  121. status = ql_read_other_func_serdes_reg(qdev,
  122. XG_SERDES_XAUI_HSS_PCS_START, &temp);
  123. if (status)
  124. temp = XG_SERDES_ADDR_XAUI_PWR_DOWN;
  125. if ((temp & XG_SERDES_ADDR_XAUI_PWR_DOWN) ==
  126. XG_SERDES_ADDR_XAUI_PWR_DOWN)
  127. xaui_indirect_valid = 0;
  128. status = ql_read_serdes_reg(qdev, XG_SERDES_XAUI_HSS_PCS_START, &temp);
  129. if (status)
  130. temp = XG_SERDES_ADDR_XAUI_PWR_DOWN;
  131. if ((temp & XG_SERDES_ADDR_XAUI_PWR_DOWN) ==
  132. XG_SERDES_ADDR_XAUI_PWR_DOWN)
  133. xaui_direct_valid = 0;
  134. /*
  135. * XFI register is shared so only need to read one
  136. * functions and then check the bits.
  137. */
  138. status = ql_read_serdes_reg(qdev, XG_SERDES_ADDR_STS, &temp);
  139. if (status)
  140. temp = 0;
  141. if ((temp & XG_SERDES_ADDR_XFI1_PWR_UP) ==
  142. XG_SERDES_ADDR_XFI1_PWR_UP) {
  143. /* now see if i'm NIC 1 or NIC 2 */
  144. if (qdev->func & 1)
  145. /* I'm NIC 2, so the indirect (NIC1) xfi is up. */
  146. xfi_indirect_valid = 1;
  147. else
  148. xfi_direct_valid = 1;
  149. }
  150. if ((temp & XG_SERDES_ADDR_XFI2_PWR_UP) ==
  151. XG_SERDES_ADDR_XFI2_PWR_UP) {
  152. /* now see if i'm NIC 1 or NIC 2 */
  153. if (qdev->func & 1)
  154. /* I'm NIC 2, so the indirect (NIC1) xfi is up. */
  155. xfi_direct_valid = 1;
  156. else
  157. xfi_indirect_valid = 1;
  158. }
  159. /* Get XAUI_AN register block. */
  160. if (qdev->func & 1) {
  161. /* Function 2 is direct */
  162. direct_ptr = mpi_coredump->serdes2_xaui_an;
  163. indirect_ptr = mpi_coredump->serdes_xaui_an;
  164. } else {
  165. /* Function 1 is direct */
  166. direct_ptr = mpi_coredump->serdes_xaui_an;
  167. indirect_ptr = mpi_coredump->serdes2_xaui_an;
  168. }
  169. for (i = 0; i <= 0x000000034; i += 4, direct_ptr++, indirect_ptr++)
  170. ql_get_both_serdes(qdev, i, direct_ptr, indirect_ptr,
  171. xaui_direct_valid, xaui_indirect_valid);
  172. /* Get XAUI_HSS_PCS register block. */
  173. if (qdev->func & 1) {
  174. direct_ptr =
  175. mpi_coredump->serdes2_xaui_hss_pcs;
  176. indirect_ptr =
  177. mpi_coredump->serdes_xaui_hss_pcs;
  178. } else {
  179. direct_ptr =
  180. mpi_coredump->serdes_xaui_hss_pcs;
  181. indirect_ptr =
  182. mpi_coredump->serdes2_xaui_hss_pcs;
  183. }
  184. for (i = 0x800; i <= 0x880; i += 4, direct_ptr++, indirect_ptr++)
  185. ql_get_both_serdes(qdev, i, direct_ptr, indirect_ptr,
  186. xaui_direct_valid, xaui_indirect_valid);
  187. /* Get XAUI_XFI_AN register block. */
  188. if (qdev->func & 1) {
  189. direct_ptr = mpi_coredump->serdes2_xfi_an;
  190. indirect_ptr = mpi_coredump->serdes_xfi_an;
  191. } else {
  192. direct_ptr = mpi_coredump->serdes_xfi_an;
  193. indirect_ptr = mpi_coredump->serdes2_xfi_an;
  194. }
  195. for (i = 0x1000; i <= 0x1034; i += 4, direct_ptr++, indirect_ptr++)
  196. ql_get_both_serdes(qdev, i, direct_ptr, indirect_ptr,
  197. xfi_direct_valid, xfi_indirect_valid);
  198. /* Get XAUI_XFI_TRAIN register block. */
  199. if (qdev->func & 1) {
  200. direct_ptr = mpi_coredump->serdes2_xfi_train;
  201. indirect_ptr =
  202. mpi_coredump->serdes_xfi_train;
  203. } else {
  204. direct_ptr = mpi_coredump->serdes_xfi_train;
  205. indirect_ptr =
  206. mpi_coredump->serdes2_xfi_train;
  207. }
  208. for (i = 0x1050; i <= 0x107c; i += 4, direct_ptr++, indirect_ptr++)
  209. ql_get_both_serdes(qdev, i, direct_ptr, indirect_ptr,
  210. xfi_direct_valid, xfi_indirect_valid);
  211. /* Get XAUI_XFI_HSS_PCS register block. */
  212. if (qdev->func & 1) {
  213. direct_ptr =
  214. mpi_coredump->serdes2_xfi_hss_pcs;
  215. indirect_ptr =
  216. mpi_coredump->serdes_xfi_hss_pcs;
  217. } else {
  218. direct_ptr =
  219. mpi_coredump->serdes_xfi_hss_pcs;
  220. indirect_ptr =
  221. mpi_coredump->serdes2_xfi_hss_pcs;
  222. }
  223. for (i = 0x1800; i <= 0x1838; i += 4, direct_ptr++, indirect_ptr++)
  224. ql_get_both_serdes(qdev, i, direct_ptr, indirect_ptr,
  225. xfi_direct_valid, xfi_indirect_valid);
  226. /* Get XAUI_XFI_HSS_TX register block. */
  227. if (qdev->func & 1) {
  228. direct_ptr =
  229. mpi_coredump->serdes2_xfi_hss_tx;
  230. indirect_ptr =
  231. mpi_coredump->serdes_xfi_hss_tx;
  232. } else {
  233. direct_ptr = mpi_coredump->serdes_xfi_hss_tx;
  234. indirect_ptr =
  235. mpi_coredump->serdes2_xfi_hss_tx;
  236. }
  237. for (i = 0x1c00; i <= 0x1c1f; i++, direct_ptr++, indirect_ptr++)
  238. ql_get_both_serdes(qdev, i, direct_ptr, indirect_ptr,
  239. xfi_direct_valid, xfi_indirect_valid);
  240. /* Get XAUI_XFI_HSS_RX register block. */
  241. if (qdev->func & 1) {
  242. direct_ptr =
  243. mpi_coredump->serdes2_xfi_hss_rx;
  244. indirect_ptr =
  245. mpi_coredump->serdes_xfi_hss_rx;
  246. } else {
  247. direct_ptr = mpi_coredump->serdes_xfi_hss_rx;
  248. indirect_ptr =
  249. mpi_coredump->serdes2_xfi_hss_rx;
  250. }
  251. for (i = 0x1c40; i <= 0x1c5f; i++, direct_ptr++, indirect_ptr++)
  252. ql_get_both_serdes(qdev, i, direct_ptr, indirect_ptr,
  253. xfi_direct_valid, xfi_indirect_valid);
  254. /* Get XAUI_XFI_HSS_PLL register block. */
  255. if (qdev->func & 1) {
  256. direct_ptr =
  257. mpi_coredump->serdes2_xfi_hss_pll;
  258. indirect_ptr =
  259. mpi_coredump->serdes_xfi_hss_pll;
  260. } else {
  261. direct_ptr =
  262. mpi_coredump->serdes_xfi_hss_pll;
  263. indirect_ptr =
  264. mpi_coredump->serdes2_xfi_hss_pll;
  265. }
  266. for (i = 0x1e00; i <= 0x1e1f; i++, direct_ptr++, indirect_ptr++)
  267. ql_get_both_serdes(qdev, i, direct_ptr, indirect_ptr,
  268. xfi_direct_valid, xfi_indirect_valid);
  269. return 0;
  270. }
  271. static int ql_read_other_func_xgmac_reg(struct ql_adapter *qdev, u32 reg,
  272. u32 *data)
  273. {
  274. int status = 0;
  275. /* wait for reg to come ready */
  276. status = ql_wait_other_func_reg_rdy(qdev, XGMAC_ADDR / 4,
  277. XGMAC_ADDR_RDY, XGMAC_ADDR_XME);
  278. if (status)
  279. goto exit;
  280. /* set up for reg read */
  281. ql_write_other_func_reg(qdev, XGMAC_ADDR / 4, reg | XGMAC_ADDR_R);
  282. /* wait for reg to come ready */
  283. status = ql_wait_other_func_reg_rdy(qdev, XGMAC_ADDR / 4,
  284. XGMAC_ADDR_RDY, XGMAC_ADDR_XME);
  285. if (status)
  286. goto exit;
  287. /* get the data */
  288. *data = ql_read_other_func_reg(qdev, XGMAC_DATA / 4);
  289. exit:
  290. return status;
  291. }
  292. /* Read the 400 xgmac control/statistics registers
  293. * skipping unused locations.
  294. */
  295. static int ql_get_xgmac_regs(struct ql_adapter *qdev, u32 *buf,
  296. unsigned int other_function)
  297. {
  298. int status = 0;
  299. int i;
  300. for (i = PAUSE_SRC_LO; i < XGMAC_REGISTER_END; i += 4, buf++) {
  301. /* We're reading 400 xgmac registers, but we filter out
  302. * serveral locations that are non-responsive to reads.
  303. */
  304. if ((i == 0x00000114) ||
  305. (i == 0x00000118) ||
  306. (i == 0x0000013c) ||
  307. (i == 0x00000140) ||
  308. (i > 0x00000150 && i < 0x000001fc) ||
  309. (i > 0x00000278 && i < 0x000002a0) ||
  310. (i > 0x000002c0 && i < 0x000002cf) ||
  311. (i > 0x000002dc && i < 0x000002f0) ||
  312. (i > 0x000003c8 && i < 0x00000400) ||
  313. (i > 0x00000400 && i < 0x00000410) ||
  314. (i > 0x00000410 && i < 0x00000420) ||
  315. (i > 0x00000420 && i < 0x00000430) ||
  316. (i > 0x00000430 && i < 0x00000440) ||
  317. (i > 0x00000440 && i < 0x00000450) ||
  318. (i > 0x00000450 && i < 0x00000500) ||
  319. (i > 0x0000054c && i < 0x00000568) ||
  320. (i > 0x000005c8 && i < 0x00000600)) {
  321. if (other_function)
  322. status =
  323. ql_read_other_func_xgmac_reg(qdev, i, buf);
  324. else
  325. status = ql_read_xgmac_reg(qdev, i, buf);
  326. if (status)
  327. *buf = 0xdeadbeef;
  328. break;
  329. }
  330. }
  331. return status;
  332. }
  333. static int ql_get_ets_regs(struct ql_adapter *qdev, u32 *buf)
  334. {
  335. int status = 0;
  336. int i;
  337. for (i = 0; i < 8; i++, buf++) {
  338. ql_write32(qdev, NIC_ETS, i << 29 | 0x08000000);
  339. *buf = ql_read32(qdev, NIC_ETS);
  340. }
  341. for (i = 0; i < 2; i++, buf++) {
  342. ql_write32(qdev, CNA_ETS, i << 29 | 0x08000000);
  343. *buf = ql_read32(qdev, CNA_ETS);
  344. }
  345. return status;
  346. }
  347. static void ql_get_intr_states(struct ql_adapter *qdev, u32 *buf)
  348. {
  349. int i;
  350. for (i = 0; i < qdev->rx_ring_count; i++, buf++) {
  351. ql_write32(qdev, INTR_EN,
  352. qdev->intr_context[i].intr_read_mask);
  353. *buf = ql_read32(qdev, INTR_EN);
  354. }
  355. }
  356. static int ql_get_cam_entries(struct ql_adapter *qdev, u32 *buf)
  357. {
  358. int i, status;
  359. u32 value[3];
  360. status = ql_sem_spinlock(qdev, SEM_MAC_ADDR_MASK);
  361. if (status)
  362. return status;
  363. for (i = 0; i < 16; i++) {
  364. status = ql_get_mac_addr_reg(qdev,
  365. MAC_ADDR_TYPE_CAM_MAC, i, value);
  366. if (status) {
  367. netif_err(qdev, drv, qdev->ndev,
  368. "Failed read of mac index register\n");
  369. goto err;
  370. }
  371. *buf++ = value[0]; /* lower MAC address */
  372. *buf++ = value[1]; /* upper MAC address */
  373. *buf++ = value[2]; /* output */
  374. }
  375. for (i = 0; i < 32; i++) {
  376. status = ql_get_mac_addr_reg(qdev,
  377. MAC_ADDR_TYPE_MULTI_MAC, i, value);
  378. if (status) {
  379. netif_err(qdev, drv, qdev->ndev,
  380. "Failed read of mac index register\n");
  381. goto err;
  382. }
  383. *buf++ = value[0]; /* lower Mcast address */
  384. *buf++ = value[1]; /* upper Mcast address */
  385. }
  386. err:
  387. ql_sem_unlock(qdev, SEM_MAC_ADDR_MASK);
  388. return status;
  389. }
  390. static int ql_get_routing_entries(struct ql_adapter *qdev, u32 *buf)
  391. {
  392. int status;
  393. u32 value, i;
  394. status = ql_sem_spinlock(qdev, SEM_RT_IDX_MASK);
  395. if (status)
  396. return status;
  397. for (i = 0; i < 16; i++) {
  398. status = ql_get_routing_reg(qdev, i, &value);
  399. if (status) {
  400. netif_err(qdev, drv, qdev->ndev,
  401. "Failed read of routing index register\n");
  402. goto err;
  403. } else {
  404. *buf++ = value;
  405. }
  406. }
  407. err:
  408. ql_sem_unlock(qdev, SEM_RT_IDX_MASK);
  409. return status;
  410. }
  411. /* Read the MPI Processor shadow registers */
  412. static int ql_get_mpi_shadow_regs(struct ql_adapter *qdev, u32 *buf)
  413. {
  414. u32 i;
  415. int status;
  416. for (i = 0; i < MPI_CORE_SH_REGS_CNT; i++, buf++) {
  417. status = ql_write_mpi_reg(qdev, RISC_124,
  418. (SHADOW_OFFSET | i << SHADOW_REG_SHIFT));
  419. if (status)
  420. goto end;
  421. status = ql_read_mpi_reg(qdev, RISC_127, buf);
  422. if (status)
  423. goto end;
  424. }
  425. end:
  426. return status;
  427. }
  428. /* Read the MPI Processor core registers */
  429. static int ql_get_mpi_regs(struct ql_adapter *qdev, u32 *buf,
  430. u32 offset, u32 count)
  431. {
  432. int i, status = 0;
  433. for (i = 0; i < count; i++, buf++) {
  434. status = ql_read_mpi_reg(qdev, offset + i, buf);
  435. if (status)
  436. return status;
  437. }
  438. return status;
  439. }
  440. /* Read the ASIC probe dump */
  441. static unsigned int *ql_get_probe(struct ql_adapter *qdev, u32 clock,
  442. u32 valid, u32 *buf)
  443. {
  444. u32 module, mux_sel, probe, lo_val, hi_val;
  445. for (module = 0; module < PRB_MX_ADDR_MAX_MODS; module++) {
  446. if (!((valid >> module) & 1))
  447. continue;
  448. for (mux_sel = 0; mux_sel < PRB_MX_ADDR_MAX_MUX; mux_sel++) {
  449. probe = clock
  450. | PRB_MX_ADDR_ARE
  451. | mux_sel
  452. | (module << PRB_MX_ADDR_MOD_SEL_SHIFT);
  453. ql_write32(qdev, PRB_MX_ADDR, probe);
  454. lo_val = ql_read32(qdev, PRB_MX_DATA);
  455. if (mux_sel == 0) {
  456. *buf = probe;
  457. buf++;
  458. }
  459. probe |= PRB_MX_ADDR_UP;
  460. ql_write32(qdev, PRB_MX_ADDR, probe);
  461. hi_val = ql_read32(qdev, PRB_MX_DATA);
  462. *buf = lo_val;
  463. buf++;
  464. *buf = hi_val;
  465. buf++;
  466. }
  467. }
  468. return buf;
  469. }
  470. static int ql_get_probe_dump(struct ql_adapter *qdev, unsigned int *buf)
  471. {
  472. /* First we have to enable the probe mux */
  473. ql_write_mpi_reg(qdev, MPI_TEST_FUNC_PRB_CTL, MPI_TEST_FUNC_PRB_EN);
  474. buf = ql_get_probe(qdev, PRB_MX_ADDR_SYS_CLOCK,
  475. PRB_MX_ADDR_VALID_SYS_MOD, buf);
  476. buf = ql_get_probe(qdev, PRB_MX_ADDR_PCI_CLOCK,
  477. PRB_MX_ADDR_VALID_PCI_MOD, buf);
  478. buf = ql_get_probe(qdev, PRB_MX_ADDR_XGM_CLOCK,
  479. PRB_MX_ADDR_VALID_XGM_MOD, buf);
  480. buf = ql_get_probe(qdev, PRB_MX_ADDR_FC_CLOCK,
  481. PRB_MX_ADDR_VALID_FC_MOD, buf);
  482. return 0;
  483. }
  484. /* Read out the routing index registers */
  485. static int ql_get_routing_index_registers(struct ql_adapter *qdev, u32 *buf)
  486. {
  487. int status;
  488. u32 type, index, index_max;
  489. u32 result_index;
  490. u32 result_data;
  491. u32 val;
  492. status = ql_sem_spinlock(qdev, SEM_RT_IDX_MASK);
  493. if (status)
  494. return status;
  495. for (type = 0; type < 4; type++) {
  496. if (type < 2)
  497. index_max = 8;
  498. else
  499. index_max = 16;
  500. for (index = 0; index < index_max; index++) {
  501. val = RT_IDX_RS
  502. | (type << RT_IDX_TYPE_SHIFT)
  503. | (index << RT_IDX_IDX_SHIFT);
  504. ql_write32(qdev, RT_IDX, val);
  505. result_index = 0;
  506. while ((result_index & RT_IDX_MR) == 0)
  507. result_index = ql_read32(qdev, RT_IDX);
  508. result_data = ql_read32(qdev, RT_DATA);
  509. *buf = type;
  510. buf++;
  511. *buf = index;
  512. buf++;
  513. *buf = result_index;
  514. buf++;
  515. *buf = result_data;
  516. buf++;
  517. }
  518. }
  519. ql_sem_unlock(qdev, SEM_RT_IDX_MASK);
  520. return status;
  521. }
  522. /* Read out the MAC protocol registers */
  523. static void ql_get_mac_protocol_registers(struct ql_adapter *qdev, u32 *buf)
  524. {
  525. u32 result_index, result_data;
  526. u32 type;
  527. u32 index;
  528. u32 offset;
  529. u32 val;
  530. u32 initial_val = MAC_ADDR_RS;
  531. u32 max_index;
  532. u32 max_offset;
  533. for (type = 0; type < MAC_ADDR_TYPE_COUNT; type++) {
  534. switch (type) {
  535. case 0: /* CAM */
  536. initial_val |= MAC_ADDR_ADR;
  537. max_index = MAC_ADDR_MAX_CAM_ENTRIES;
  538. max_offset = MAC_ADDR_MAX_CAM_WCOUNT;
  539. break;
  540. case 1: /* Multicast MAC Address */
  541. max_index = MAC_ADDR_MAX_CAM_WCOUNT;
  542. max_offset = MAC_ADDR_MAX_CAM_WCOUNT;
  543. break;
  544. case 2: /* VLAN filter mask */
  545. case 3: /* MC filter mask */
  546. max_index = MAC_ADDR_MAX_CAM_WCOUNT;
  547. max_offset = MAC_ADDR_MAX_CAM_WCOUNT;
  548. break;
  549. case 4: /* FC MAC addresses */
  550. max_index = MAC_ADDR_MAX_FC_MAC_ENTRIES;
  551. max_offset = MAC_ADDR_MAX_FC_MAC_WCOUNT;
  552. break;
  553. case 5: /* Mgmt MAC addresses */
  554. max_index = MAC_ADDR_MAX_MGMT_MAC_ENTRIES;
  555. max_offset = MAC_ADDR_MAX_MGMT_MAC_WCOUNT;
  556. break;
  557. case 6: /* Mgmt VLAN addresses */
  558. max_index = MAC_ADDR_MAX_MGMT_VLAN_ENTRIES;
  559. max_offset = MAC_ADDR_MAX_MGMT_VLAN_WCOUNT;
  560. break;
  561. case 7: /* Mgmt IPv4 address */
  562. max_index = MAC_ADDR_MAX_MGMT_V4_ENTRIES;
  563. max_offset = MAC_ADDR_MAX_MGMT_V4_WCOUNT;
  564. break;
  565. case 8: /* Mgmt IPv6 address */
  566. max_index = MAC_ADDR_MAX_MGMT_V6_ENTRIES;
  567. max_offset = MAC_ADDR_MAX_MGMT_V6_WCOUNT;
  568. break;
  569. case 9: /* Mgmt TCP/UDP Dest port */
  570. max_index = MAC_ADDR_MAX_MGMT_TU_DP_ENTRIES;
  571. max_offset = MAC_ADDR_MAX_MGMT_TU_DP_WCOUNT;
  572. break;
  573. default:
  574. pr_err("Bad type!!! 0x%08x\n", type);
  575. max_index = 0;
  576. max_offset = 0;
  577. break;
  578. }
  579. for (index = 0; index < max_index; index++) {
  580. for (offset = 0; offset < max_offset; offset++) {
  581. val = initial_val
  582. | (type << MAC_ADDR_TYPE_SHIFT)
  583. | (index << MAC_ADDR_IDX_SHIFT)
  584. | (offset);
  585. ql_write32(qdev, MAC_ADDR_IDX, val);
  586. result_index = 0;
  587. while ((result_index & MAC_ADDR_MR) == 0) {
  588. result_index = ql_read32(qdev,
  589. MAC_ADDR_IDX);
  590. }
  591. result_data = ql_read32(qdev, MAC_ADDR_DATA);
  592. *buf = result_index;
  593. buf++;
  594. *buf = result_data;
  595. buf++;
  596. }
  597. }
  598. }
  599. }
  600. static void ql_get_sem_registers(struct ql_adapter *qdev, u32 *buf)
  601. {
  602. u32 func_num, reg, reg_val;
  603. int status;
  604. for (func_num = 0; func_num < MAX_SEMAPHORE_FUNCTIONS ; func_num++) {
  605. reg = MPI_NIC_REG_BLOCK
  606. | (func_num << MPI_NIC_FUNCTION_SHIFT)
  607. | (SEM / 4);
  608. status = ql_read_mpi_reg(qdev, reg, &reg_val);
  609. *buf = reg_val;
  610. /* if the read failed then dead fill the element. */
  611. if (!status)
  612. *buf = 0xdeadbeef;
  613. buf++;
  614. }
  615. }
  616. /* Create a coredump segment header */
  617. static void ql_build_coredump_seg_header(
  618. struct mpi_coredump_segment_header *seg_hdr,
  619. u32 seg_number, u32 seg_size, u8 *desc)
  620. {
  621. memset(seg_hdr, 0, sizeof(struct mpi_coredump_segment_header));
  622. seg_hdr->cookie = MPI_COREDUMP_COOKIE;
  623. seg_hdr->segNum = seg_number;
  624. seg_hdr->segSize = seg_size;
  625. strncpy(seg_hdr->description, desc, (sizeof(seg_hdr->description)) - 1);
  626. }
  627. /*
  628. * This function should be called when a coredump / probedump
  629. * is to be extracted from the HBA. It is assumed there is a
  630. * qdev structure that contains the base address of the register
  631. * space for this function as well as a coredump structure that
  632. * will contain the dump.
  633. */
  634. int ql_core_dump(struct ql_adapter *qdev, struct ql_mpi_coredump *mpi_coredump)
  635. {
  636. int status;
  637. int i;
  638. if (!mpi_coredump) {
  639. netif_err(qdev, drv, qdev->ndev, "No memory allocated\n");
  640. return -EINVAL;
  641. }
  642. /* Try to get the spinlock, but dont worry if
  643. * it isn't available. If the firmware died it
  644. * might be holding the sem.
  645. */
  646. ql_sem_spinlock(qdev, SEM_PROC_REG_MASK);
  647. status = ql_pause_mpi_risc(qdev);
  648. if (status) {
  649. netif_err(qdev, drv, qdev->ndev,
  650. "Failed RISC pause. Status = 0x%.08x\n", status);
  651. goto err;
  652. }
  653. /* Insert the global header */
  654. memset(&(mpi_coredump->mpi_global_header), 0,
  655. sizeof(struct mpi_coredump_global_header));
  656. mpi_coredump->mpi_global_header.cookie = MPI_COREDUMP_COOKIE;
  657. mpi_coredump->mpi_global_header.headerSize =
  658. sizeof(struct mpi_coredump_global_header);
  659. mpi_coredump->mpi_global_header.imageSize =
  660. sizeof(struct ql_mpi_coredump);
  661. strncpy(mpi_coredump->mpi_global_header.idString, "MPI Coredump",
  662. sizeof(mpi_coredump->mpi_global_header.idString));
  663. /* Get generic NIC reg dump */
  664. ql_build_coredump_seg_header(&mpi_coredump->nic_regs_seg_hdr,
  665. NIC1_CONTROL_SEG_NUM,
  666. sizeof(struct mpi_coredump_segment_header) +
  667. sizeof(mpi_coredump->nic_regs), "NIC1 Registers");
  668. ql_build_coredump_seg_header(&mpi_coredump->nic2_regs_seg_hdr,
  669. NIC2_CONTROL_SEG_NUM,
  670. sizeof(struct mpi_coredump_segment_header) +
  671. sizeof(mpi_coredump->nic2_regs), "NIC2 Registers");
  672. /* Get XGMac registers. (Segment 18, Rev C. step 21) */
  673. ql_build_coredump_seg_header(&mpi_coredump->xgmac1_seg_hdr,
  674. NIC1_XGMAC_SEG_NUM,
  675. sizeof(struct mpi_coredump_segment_header) +
  676. sizeof(mpi_coredump->xgmac1), "NIC1 XGMac Registers");
  677. ql_build_coredump_seg_header(&mpi_coredump->xgmac2_seg_hdr,
  678. NIC2_XGMAC_SEG_NUM,
  679. sizeof(struct mpi_coredump_segment_header) +
  680. sizeof(mpi_coredump->xgmac2), "NIC2 XGMac Registers");
  681. if (qdev->func & 1) {
  682. /* Odd means our function is NIC 2 */
  683. for (i = 0; i < NIC_REGS_DUMP_WORD_COUNT; i++)
  684. mpi_coredump->nic2_regs[i] =
  685. ql_read32(qdev, i * sizeof(u32));
  686. for (i = 0; i < NIC_REGS_DUMP_WORD_COUNT; i++)
  687. mpi_coredump->nic_regs[i] =
  688. ql_read_other_func_reg(qdev, (i * sizeof(u32)) / 4);
  689. ql_get_xgmac_regs(qdev, &mpi_coredump->xgmac2[0], 0);
  690. ql_get_xgmac_regs(qdev, &mpi_coredump->xgmac1[0], 1);
  691. } else {
  692. /* Even means our function is NIC 1 */
  693. for (i = 0; i < NIC_REGS_DUMP_WORD_COUNT; i++)
  694. mpi_coredump->nic_regs[i] =
  695. ql_read32(qdev, i * sizeof(u32));
  696. for (i = 0; i < NIC_REGS_DUMP_WORD_COUNT; i++)
  697. mpi_coredump->nic2_regs[i] =
  698. ql_read_other_func_reg(qdev, (i * sizeof(u32)) / 4);
  699. ql_get_xgmac_regs(qdev, &mpi_coredump->xgmac1[0], 0);
  700. ql_get_xgmac_regs(qdev, &mpi_coredump->xgmac2[0], 1);
  701. }
  702. /* Rev C. Step 20a */
  703. ql_build_coredump_seg_header(&mpi_coredump->xaui_an_hdr,
  704. XAUI_AN_SEG_NUM,
  705. sizeof(struct mpi_coredump_segment_header) +
  706. sizeof(mpi_coredump->serdes_xaui_an),
  707. "XAUI AN Registers");
  708. /* Rev C. Step 20b */
  709. ql_build_coredump_seg_header(&mpi_coredump->xaui_hss_pcs_hdr,
  710. XAUI_HSS_PCS_SEG_NUM,
  711. sizeof(struct mpi_coredump_segment_header) +
  712. sizeof(mpi_coredump->serdes_xaui_hss_pcs),
  713. "XAUI HSS PCS Registers");
  714. ql_build_coredump_seg_header(&mpi_coredump->xfi_an_hdr, XFI_AN_SEG_NUM,
  715. sizeof(struct mpi_coredump_segment_header) +
  716. sizeof(mpi_coredump->serdes_xfi_an),
  717. "XFI AN Registers");
  718. ql_build_coredump_seg_header(&mpi_coredump->xfi_train_hdr,
  719. XFI_TRAIN_SEG_NUM,
  720. sizeof(struct mpi_coredump_segment_header) +
  721. sizeof(mpi_coredump->serdes_xfi_train),
  722. "XFI TRAIN Registers");
  723. ql_build_coredump_seg_header(&mpi_coredump->xfi_hss_pcs_hdr,
  724. XFI_HSS_PCS_SEG_NUM,
  725. sizeof(struct mpi_coredump_segment_header) +
  726. sizeof(mpi_coredump->serdes_xfi_hss_pcs),
  727. "XFI HSS PCS Registers");
  728. ql_build_coredump_seg_header(&mpi_coredump->xfi_hss_tx_hdr,
  729. XFI_HSS_TX_SEG_NUM,
  730. sizeof(struct mpi_coredump_segment_header) +
  731. sizeof(mpi_coredump->serdes_xfi_hss_tx),
  732. "XFI HSS TX Registers");
  733. ql_build_coredump_seg_header(&mpi_coredump->xfi_hss_rx_hdr,
  734. XFI_HSS_RX_SEG_NUM,
  735. sizeof(struct mpi_coredump_segment_header) +
  736. sizeof(mpi_coredump->serdes_xfi_hss_rx),
  737. "XFI HSS RX Registers");
  738. ql_build_coredump_seg_header(&mpi_coredump->xfi_hss_pll_hdr,
  739. XFI_HSS_PLL_SEG_NUM,
  740. sizeof(struct mpi_coredump_segment_header) +
  741. sizeof(mpi_coredump->serdes_xfi_hss_pll),
  742. "XFI HSS PLL Registers");
  743. ql_build_coredump_seg_header(&mpi_coredump->xaui2_an_hdr,
  744. XAUI2_AN_SEG_NUM,
  745. sizeof(struct mpi_coredump_segment_header) +
  746. sizeof(mpi_coredump->serdes2_xaui_an),
  747. "XAUI2 AN Registers");
  748. ql_build_coredump_seg_header(&mpi_coredump->xaui2_hss_pcs_hdr,
  749. XAUI2_HSS_PCS_SEG_NUM,
  750. sizeof(struct mpi_coredump_segment_header) +
  751. sizeof(mpi_coredump->serdes2_xaui_hss_pcs),
  752. "XAUI2 HSS PCS Registers");
  753. ql_build_coredump_seg_header(&mpi_coredump->xfi2_an_hdr,
  754. XFI2_AN_SEG_NUM,
  755. sizeof(struct mpi_coredump_segment_header) +
  756. sizeof(mpi_coredump->serdes2_xfi_an),
  757. "XFI2 AN Registers");
  758. ql_build_coredump_seg_header(&mpi_coredump->xfi2_train_hdr,
  759. XFI2_TRAIN_SEG_NUM,
  760. sizeof(struct mpi_coredump_segment_header) +
  761. sizeof(mpi_coredump->serdes2_xfi_train),
  762. "XFI2 TRAIN Registers");
  763. ql_build_coredump_seg_header(&mpi_coredump->xfi2_hss_pcs_hdr,
  764. XFI2_HSS_PCS_SEG_NUM,
  765. sizeof(struct mpi_coredump_segment_header) +
  766. sizeof(mpi_coredump->serdes2_xfi_hss_pcs),
  767. "XFI2 HSS PCS Registers");
  768. ql_build_coredump_seg_header(&mpi_coredump->xfi2_hss_tx_hdr,
  769. XFI2_HSS_TX_SEG_NUM,
  770. sizeof(struct mpi_coredump_segment_header) +
  771. sizeof(mpi_coredump->serdes2_xfi_hss_tx),
  772. "XFI2 HSS TX Registers");
  773. ql_build_coredump_seg_header(&mpi_coredump->xfi2_hss_rx_hdr,
  774. XFI2_HSS_RX_SEG_NUM,
  775. sizeof(struct mpi_coredump_segment_header) +
  776. sizeof(mpi_coredump->serdes2_xfi_hss_rx),
  777. "XFI2 HSS RX Registers");
  778. ql_build_coredump_seg_header(&mpi_coredump->xfi2_hss_pll_hdr,
  779. XFI2_HSS_PLL_SEG_NUM,
  780. sizeof(struct mpi_coredump_segment_header) +
  781. sizeof(mpi_coredump->serdes2_xfi_hss_pll),
  782. "XFI2 HSS PLL Registers");
  783. status = ql_get_serdes_regs(qdev, mpi_coredump);
  784. if (status) {
  785. netif_err(qdev, drv, qdev->ndev,
  786. "Failed Dump of Serdes Registers. Status = 0x%.08x\n",
  787. status);
  788. goto err;
  789. }
  790. ql_build_coredump_seg_header(&mpi_coredump->core_regs_seg_hdr,
  791. CORE_SEG_NUM,
  792. sizeof(mpi_coredump->core_regs_seg_hdr) +
  793. sizeof(mpi_coredump->mpi_core_regs) +
  794. sizeof(mpi_coredump->mpi_core_sh_regs),
  795. "Core Registers");
  796. /* Get the MPI Core Registers */
  797. status = ql_get_mpi_regs(qdev, &mpi_coredump->mpi_core_regs[0],
  798. MPI_CORE_REGS_ADDR, MPI_CORE_REGS_CNT);
  799. if (status)
  800. goto err;
  801. /* Get the 16 MPI shadow registers */
  802. status = ql_get_mpi_shadow_regs(qdev,
  803. &mpi_coredump->mpi_core_sh_regs[0]);
  804. if (status)
  805. goto err;
  806. /* Get the Test Logic Registers */
  807. ql_build_coredump_seg_header(&mpi_coredump->test_logic_regs_seg_hdr,
  808. TEST_LOGIC_SEG_NUM,
  809. sizeof(struct mpi_coredump_segment_header)
  810. + sizeof(mpi_coredump->test_logic_regs),
  811. "Test Logic Regs");
  812. status = ql_get_mpi_regs(qdev, &mpi_coredump->test_logic_regs[0],
  813. TEST_REGS_ADDR, TEST_REGS_CNT);
  814. if (status)
  815. goto err;
  816. /* Get the RMII Registers */
  817. ql_build_coredump_seg_header(&mpi_coredump->rmii_regs_seg_hdr,
  818. RMII_SEG_NUM,
  819. sizeof(struct mpi_coredump_segment_header)
  820. + sizeof(mpi_coredump->rmii_regs),
  821. "RMII Registers");
  822. status = ql_get_mpi_regs(qdev, &mpi_coredump->rmii_regs[0],
  823. RMII_REGS_ADDR, RMII_REGS_CNT);
  824. if (status)
  825. goto err;
  826. /* Get the FCMAC1 Registers */
  827. ql_build_coredump_seg_header(&mpi_coredump->fcmac1_regs_seg_hdr,
  828. FCMAC1_SEG_NUM,
  829. sizeof(struct mpi_coredump_segment_header)
  830. + sizeof(mpi_coredump->fcmac1_regs),
  831. "FCMAC1 Registers");
  832. status = ql_get_mpi_regs(qdev, &mpi_coredump->fcmac1_regs[0],
  833. FCMAC1_REGS_ADDR, FCMAC_REGS_CNT);
  834. if (status)
  835. goto err;
  836. /* Get the FCMAC2 Registers */
  837. ql_build_coredump_seg_header(&mpi_coredump->fcmac2_regs_seg_hdr,
  838. FCMAC2_SEG_NUM,
  839. sizeof(struct mpi_coredump_segment_header)
  840. + sizeof(mpi_coredump->fcmac2_regs),
  841. "FCMAC2 Registers");
  842. status = ql_get_mpi_regs(qdev, &mpi_coredump->fcmac2_regs[0],
  843. FCMAC2_REGS_ADDR, FCMAC_REGS_CNT);
  844. if (status)
  845. goto err;
  846. /* Get the FC1 MBX Registers */
  847. ql_build_coredump_seg_header(&mpi_coredump->fc1_mbx_regs_seg_hdr,
  848. FC1_MBOX_SEG_NUM,
  849. sizeof(struct mpi_coredump_segment_header)
  850. + sizeof(mpi_coredump->fc1_mbx_regs),
  851. "FC1 MBox Regs");
  852. status = ql_get_mpi_regs(qdev, &mpi_coredump->fc1_mbx_regs[0],
  853. FC1_MBX_REGS_ADDR, FC_MBX_REGS_CNT);
  854. if (status)
  855. goto err;
  856. /* Get the IDE Registers */
  857. ql_build_coredump_seg_header(&mpi_coredump->ide_regs_seg_hdr,
  858. IDE_SEG_NUM,
  859. sizeof(struct mpi_coredump_segment_header)
  860. + sizeof(mpi_coredump->ide_regs),
  861. "IDE Registers");
  862. status = ql_get_mpi_regs(qdev, &mpi_coredump->ide_regs[0],
  863. IDE_REGS_ADDR, IDE_REGS_CNT);
  864. if (status)
  865. goto err;
  866. /* Get the NIC1 MBX Registers */
  867. ql_build_coredump_seg_header(&mpi_coredump->nic1_mbx_regs_seg_hdr,
  868. NIC1_MBOX_SEG_NUM,
  869. sizeof(struct mpi_coredump_segment_header)
  870. + sizeof(mpi_coredump->nic1_mbx_regs),
  871. "NIC1 MBox Regs");
  872. status = ql_get_mpi_regs(qdev, &mpi_coredump->nic1_mbx_regs[0],
  873. NIC1_MBX_REGS_ADDR, NIC_MBX_REGS_CNT);
  874. if (status)
  875. goto err;
  876. /* Get the SMBus Registers */
  877. ql_build_coredump_seg_header(&mpi_coredump->smbus_regs_seg_hdr,
  878. SMBUS_SEG_NUM,
  879. sizeof(struct mpi_coredump_segment_header)
  880. + sizeof(mpi_coredump->smbus_regs),
  881. "SMBus Registers");
  882. status = ql_get_mpi_regs(qdev, &mpi_coredump->smbus_regs[0],
  883. SMBUS_REGS_ADDR, SMBUS_REGS_CNT);
  884. if (status)
  885. goto err;
  886. /* Get the FC2 MBX Registers */
  887. ql_build_coredump_seg_header(&mpi_coredump->fc2_mbx_regs_seg_hdr,
  888. FC2_MBOX_SEG_NUM,
  889. sizeof(struct mpi_coredump_segment_header)
  890. + sizeof(mpi_coredump->fc2_mbx_regs),
  891. "FC2 MBox Regs");
  892. status = ql_get_mpi_regs(qdev, &mpi_coredump->fc2_mbx_regs[0],
  893. FC2_MBX_REGS_ADDR, FC_MBX_REGS_CNT);
  894. if (status)
  895. goto err;
  896. /* Get the NIC2 MBX Registers */
  897. ql_build_coredump_seg_header(&mpi_coredump->nic2_mbx_regs_seg_hdr,
  898. NIC2_MBOX_SEG_NUM,
  899. sizeof(struct mpi_coredump_segment_header)
  900. + sizeof(mpi_coredump->nic2_mbx_regs),
  901. "NIC2 MBox Regs");
  902. status = ql_get_mpi_regs(qdev, &mpi_coredump->nic2_mbx_regs[0],
  903. NIC2_MBX_REGS_ADDR, NIC_MBX_REGS_CNT);
  904. if (status)
  905. goto err;
  906. /* Get the I2C Registers */
  907. ql_build_coredump_seg_header(&mpi_coredump->i2c_regs_seg_hdr,
  908. I2C_SEG_NUM,
  909. sizeof(struct mpi_coredump_segment_header)
  910. + sizeof(mpi_coredump->i2c_regs),
  911. "I2C Registers");
  912. status = ql_get_mpi_regs(qdev, &mpi_coredump->i2c_regs[0],
  913. I2C_REGS_ADDR, I2C_REGS_CNT);
  914. if (status)
  915. goto err;
  916. /* Get the MEMC Registers */
  917. ql_build_coredump_seg_header(&mpi_coredump->memc_regs_seg_hdr,
  918. MEMC_SEG_NUM,
  919. sizeof(struct mpi_coredump_segment_header)
  920. + sizeof(mpi_coredump->memc_regs),
  921. "MEMC Registers");
  922. status = ql_get_mpi_regs(qdev, &mpi_coredump->memc_regs[0],
  923. MEMC_REGS_ADDR, MEMC_REGS_CNT);
  924. if (status)
  925. goto err;
  926. /* Get the PBus Registers */
  927. ql_build_coredump_seg_header(&mpi_coredump->pbus_regs_seg_hdr,
  928. PBUS_SEG_NUM,
  929. sizeof(struct mpi_coredump_segment_header)
  930. + sizeof(mpi_coredump->pbus_regs),
  931. "PBUS Registers");
  932. status = ql_get_mpi_regs(qdev, &mpi_coredump->pbus_regs[0],
  933. PBUS_REGS_ADDR, PBUS_REGS_CNT);
  934. if (status)
  935. goto err;
  936. /* Get the MDE Registers */
  937. ql_build_coredump_seg_header(&mpi_coredump->mde_regs_seg_hdr,
  938. MDE_SEG_NUM,
  939. sizeof(struct mpi_coredump_segment_header)
  940. + sizeof(mpi_coredump->mde_regs),
  941. "MDE Registers");
  942. status = ql_get_mpi_regs(qdev, &mpi_coredump->mde_regs[0],
  943. MDE_REGS_ADDR, MDE_REGS_CNT);
  944. if (status)
  945. goto err;
  946. ql_build_coredump_seg_header(&mpi_coredump->misc_nic_seg_hdr,
  947. MISC_NIC_INFO_SEG_NUM,
  948. sizeof(struct mpi_coredump_segment_header)
  949. + sizeof(mpi_coredump->misc_nic_info),
  950. "MISC NIC INFO");
  951. mpi_coredump->misc_nic_info.rx_ring_count = qdev->rx_ring_count;
  952. mpi_coredump->misc_nic_info.tx_ring_count = qdev->tx_ring_count;
  953. mpi_coredump->misc_nic_info.intr_count = qdev->intr_count;
  954. mpi_coredump->misc_nic_info.function = qdev->func;
  955. /* Segment 31 */
  956. /* Get indexed register values. */
  957. ql_build_coredump_seg_header(&mpi_coredump->intr_states_seg_hdr,
  958. INTR_STATES_SEG_NUM,
  959. sizeof(struct mpi_coredump_segment_header)
  960. + sizeof(mpi_coredump->intr_states),
  961. "INTR States");
  962. ql_get_intr_states(qdev, &mpi_coredump->intr_states[0]);
  963. ql_build_coredump_seg_header(&mpi_coredump->cam_entries_seg_hdr,
  964. CAM_ENTRIES_SEG_NUM,
  965. sizeof(struct mpi_coredump_segment_header)
  966. + sizeof(mpi_coredump->cam_entries),
  967. "CAM Entries");
  968. status = ql_get_cam_entries(qdev, &mpi_coredump->cam_entries[0]);
  969. if (status)
  970. goto err;
  971. ql_build_coredump_seg_header(&mpi_coredump->nic_routing_words_seg_hdr,
  972. ROUTING_WORDS_SEG_NUM,
  973. sizeof(struct mpi_coredump_segment_header)
  974. + sizeof(mpi_coredump->nic_routing_words),
  975. "Routing Words");
  976. status = ql_get_routing_entries(qdev,
  977. &mpi_coredump->nic_routing_words[0]);
  978. if (status)
  979. goto err;
  980. /* Segment 34 (Rev C. step 23) */
  981. ql_build_coredump_seg_header(&mpi_coredump->ets_seg_hdr,
  982. ETS_SEG_NUM,
  983. sizeof(struct mpi_coredump_segment_header)
  984. + sizeof(mpi_coredump->ets),
  985. "ETS Registers");
  986. status = ql_get_ets_regs(qdev, &mpi_coredump->ets[0]);
  987. if (status)
  988. goto err;
  989. ql_build_coredump_seg_header(&mpi_coredump->probe_dump_seg_hdr,
  990. PROBE_DUMP_SEG_NUM,
  991. sizeof(struct mpi_coredump_segment_header)
  992. + sizeof(mpi_coredump->probe_dump),
  993. "Probe Dump");
  994. ql_get_probe_dump(qdev, &mpi_coredump->probe_dump[0]);
  995. ql_build_coredump_seg_header(&mpi_coredump->routing_reg_seg_hdr,
  996. ROUTING_INDEX_SEG_NUM,
  997. sizeof(struct mpi_coredump_segment_header)
  998. + sizeof(mpi_coredump->routing_regs),
  999. "Routing Regs");
  1000. status = ql_get_routing_index_registers(qdev,
  1001. &mpi_coredump->routing_regs[0]);
  1002. if (status)
  1003. goto err;
  1004. ql_build_coredump_seg_header(&mpi_coredump->mac_prot_reg_seg_hdr,
  1005. MAC_PROTOCOL_SEG_NUM,
  1006. sizeof(struct mpi_coredump_segment_header)
  1007. + sizeof(mpi_coredump->mac_prot_regs),
  1008. "MAC Prot Regs");
  1009. ql_get_mac_protocol_registers(qdev, &mpi_coredump->mac_prot_regs[0]);
  1010. /* Get the semaphore registers for all 5 functions */
  1011. ql_build_coredump_seg_header(&mpi_coredump->sem_regs_seg_hdr,
  1012. SEM_REGS_SEG_NUM,
  1013. sizeof(struct mpi_coredump_segment_header) +
  1014. sizeof(mpi_coredump->sem_regs), "Sem Registers");
  1015. ql_get_sem_registers(qdev, &mpi_coredump->sem_regs[0]);
  1016. /* Prevent the mpi restarting while we dump the memory.*/
  1017. ql_write_mpi_reg(qdev, MPI_TEST_FUNC_RST_STS, MPI_TEST_FUNC_RST_FRC);
  1018. /* clear the pause */
  1019. status = ql_unpause_mpi_risc(qdev);
  1020. if (status) {
  1021. netif_err(qdev, drv, qdev->ndev,
  1022. "Failed RISC unpause. Status = 0x%.08x\n", status);
  1023. goto err;
  1024. }
  1025. /* Reset the RISC so we can dump RAM */
  1026. status = ql_hard_reset_mpi_risc(qdev);
  1027. if (status) {
  1028. netif_err(qdev, drv, qdev->ndev,
  1029. "Failed RISC reset. Status = 0x%.08x\n", status);
  1030. goto err;
  1031. }
  1032. ql_build_coredump_seg_header(&mpi_coredump->code_ram_seg_hdr,
  1033. WCS_RAM_SEG_NUM,
  1034. sizeof(struct mpi_coredump_segment_header)
  1035. + sizeof(mpi_coredump->code_ram),
  1036. "WCS RAM");
  1037. status = ql_dump_risc_ram_area(qdev, &mpi_coredump->code_ram[0],
  1038. CODE_RAM_ADDR, CODE_RAM_CNT);
  1039. if (status) {
  1040. netif_err(qdev, drv, qdev->ndev,
  1041. "Failed Dump of CODE RAM. Status = 0x%.08x\n",
  1042. status);
  1043. goto err;
  1044. }
  1045. /* Insert the segment header */
  1046. ql_build_coredump_seg_header(&mpi_coredump->memc_ram_seg_hdr,
  1047. MEMC_RAM_SEG_NUM,
  1048. sizeof(struct mpi_coredump_segment_header)
  1049. + sizeof(mpi_coredump->memc_ram),
  1050. "MEMC RAM");
  1051. status = ql_dump_risc_ram_area(qdev, &mpi_coredump->memc_ram[0],
  1052. MEMC_RAM_ADDR, MEMC_RAM_CNT);
  1053. if (status) {
  1054. netif_err(qdev, drv, qdev->ndev,
  1055. "Failed Dump of MEMC RAM. Status = 0x%.08x\n",
  1056. status);
  1057. goto err;
  1058. }
  1059. err:
  1060. ql_sem_unlock(qdev, SEM_PROC_REG_MASK); /* does flush too */
  1061. return status;
  1062. }
  1063. static void ql_get_core_dump(struct ql_adapter *qdev)
  1064. {
  1065. if (!ql_own_firmware(qdev)) {
  1066. netif_err(qdev, drv, qdev->ndev, "Don't own firmware!\n");
  1067. return;
  1068. }
  1069. if (!netif_running(qdev->ndev)) {
  1070. netif_err(qdev, ifup, qdev->ndev,
  1071. "Force Coredump can only be done from interface that is up\n");
  1072. return;
  1073. }
  1074. ql_queue_fw_error(qdev);
  1075. }
  1076. static void ql_gen_reg_dump(struct ql_adapter *qdev,
  1077. struct ql_reg_dump *mpi_coredump)
  1078. {
  1079. int i, status;
  1080. memset(&(mpi_coredump->mpi_global_header), 0,
  1081. sizeof(struct mpi_coredump_global_header));
  1082. mpi_coredump->mpi_global_header.cookie = MPI_COREDUMP_COOKIE;
  1083. mpi_coredump->mpi_global_header.headerSize =
  1084. sizeof(struct mpi_coredump_global_header);
  1085. mpi_coredump->mpi_global_header.imageSize =
  1086. sizeof(struct ql_reg_dump);
  1087. strncpy(mpi_coredump->mpi_global_header.idString, "MPI Coredump",
  1088. sizeof(mpi_coredump->mpi_global_header.idString));
  1089. /* segment 16 */
  1090. ql_build_coredump_seg_header(&mpi_coredump->misc_nic_seg_hdr,
  1091. MISC_NIC_INFO_SEG_NUM,
  1092. sizeof(struct mpi_coredump_segment_header)
  1093. + sizeof(mpi_coredump->misc_nic_info),
  1094. "MISC NIC INFO");
  1095. mpi_coredump->misc_nic_info.rx_ring_count = qdev->rx_ring_count;
  1096. mpi_coredump->misc_nic_info.tx_ring_count = qdev->tx_ring_count;
  1097. mpi_coredump->misc_nic_info.intr_count = qdev->intr_count;
  1098. mpi_coredump->misc_nic_info.function = qdev->func;
  1099. /* Segment 16, Rev C. Step 18 */
  1100. ql_build_coredump_seg_header(&mpi_coredump->nic_regs_seg_hdr,
  1101. NIC1_CONTROL_SEG_NUM,
  1102. sizeof(struct mpi_coredump_segment_header)
  1103. + sizeof(mpi_coredump->nic_regs),
  1104. "NIC Registers");
  1105. /* Get generic reg dump */
  1106. for (i = 0; i < 64; i++)
  1107. mpi_coredump->nic_regs[i] = ql_read32(qdev, i * sizeof(u32));
  1108. /* Segment 31 */
  1109. /* Get indexed register values. */
  1110. ql_build_coredump_seg_header(&mpi_coredump->intr_states_seg_hdr,
  1111. INTR_STATES_SEG_NUM,
  1112. sizeof(struct mpi_coredump_segment_header)
  1113. + sizeof(mpi_coredump->intr_states),
  1114. "INTR States");
  1115. ql_get_intr_states(qdev, &mpi_coredump->intr_states[0]);
  1116. ql_build_coredump_seg_header(&mpi_coredump->cam_entries_seg_hdr,
  1117. CAM_ENTRIES_SEG_NUM,
  1118. sizeof(struct mpi_coredump_segment_header)
  1119. + sizeof(mpi_coredump->cam_entries),
  1120. "CAM Entries");
  1121. status = ql_get_cam_entries(qdev, &mpi_coredump->cam_entries[0]);
  1122. if (status)
  1123. return;
  1124. ql_build_coredump_seg_header(&mpi_coredump->nic_routing_words_seg_hdr,
  1125. ROUTING_WORDS_SEG_NUM,
  1126. sizeof(struct mpi_coredump_segment_header)
  1127. + sizeof(mpi_coredump->nic_routing_words),
  1128. "Routing Words");
  1129. status = ql_get_routing_entries(qdev,
  1130. &mpi_coredump->nic_routing_words[0]);
  1131. if (status)
  1132. return;
  1133. /* Segment 34 (Rev C. step 23) */
  1134. ql_build_coredump_seg_header(&mpi_coredump->ets_seg_hdr,
  1135. ETS_SEG_NUM,
  1136. sizeof(struct mpi_coredump_segment_header)
  1137. + sizeof(mpi_coredump->ets),
  1138. "ETS Registers");
  1139. status = ql_get_ets_regs(qdev, &mpi_coredump->ets[0]);
  1140. if (status)
  1141. return;
  1142. }
  1143. void ql_get_dump(struct ql_adapter *qdev, void *buff)
  1144. {
  1145. /*
  1146. * If the dump has already been taken and is stored
  1147. * in our internal buffer and if force dump is set then
  1148. * just start the spool to dump it to the log file
  1149. * and also, take a snapshot of the general regs to
  1150. * to the user's buffer or else take complete dump
  1151. * to the user's buffer if force is not set.
  1152. */
  1153. if (!test_bit(QL_FRC_COREDUMP, &qdev->flags)) {
  1154. if (!ql_core_dump(qdev, buff))
  1155. ql_soft_reset_mpi_risc(qdev);
  1156. else
  1157. netif_err(qdev, drv, qdev->ndev, "coredump failed!\n");
  1158. } else {
  1159. ql_gen_reg_dump(qdev, buff);
  1160. ql_get_core_dump(qdev);
  1161. }
  1162. }
  1163. /* Coredump to messages log file using separate worker thread */
  1164. void ql_mpi_core_to_log(struct work_struct *work)
  1165. {
  1166. struct ql_adapter *qdev =
  1167. container_of(work, struct ql_adapter, mpi_core_to_log.work);
  1168. u32 *tmp, count;
  1169. int i;
  1170. count = sizeof(struct ql_mpi_coredump) / sizeof(u32);
  1171. tmp = (u32 *)qdev->mpi_coredump;
  1172. netif_printk(qdev, drv, KERN_DEBUG, qdev->ndev,
  1173. "Core is dumping to log file!\n");
  1174. for (i = 0; i < count; i += 8) {
  1175. pr_err("%.08x: %.08x %.08x %.08x %.08x %.08x "
  1176. "%.08x %.08x %.08x\n", i,
  1177. tmp[i + 0],
  1178. tmp[i + 1],
  1179. tmp[i + 2],
  1180. tmp[i + 3],
  1181. tmp[i + 4],
  1182. tmp[i + 5],
  1183. tmp[i + 6],
  1184. tmp[i + 7]);
  1185. msleep(5);
  1186. }
  1187. }
  1188. #ifdef QL_REG_DUMP
  1189. static void ql_dump_intr_states(struct ql_adapter *qdev)
  1190. {
  1191. int i;
  1192. u32 value;
  1193. for (i = 0; i < qdev->intr_count; i++) {
  1194. ql_write32(qdev, INTR_EN, qdev->intr_context[i].intr_read_mask);
  1195. value = ql_read32(qdev, INTR_EN);
  1196. pr_err("%s: Interrupt %d is %s\n",
  1197. qdev->ndev->name, i,
  1198. (value & INTR_EN_EN ? "enabled" : "disabled"));
  1199. }
  1200. }
  1201. #define DUMP_XGMAC(qdev, reg) \
  1202. do { \
  1203. u32 data; \
  1204. ql_read_xgmac_reg(qdev, reg, &data); \
  1205. pr_err("%s: %s = 0x%.08x\n", qdev->ndev->name, #reg, data); \
  1206. } while (0)
  1207. void ql_dump_xgmac_control_regs(struct ql_adapter *qdev)
  1208. {
  1209. if (ql_sem_spinlock(qdev, qdev->xg_sem_mask)) {
  1210. pr_err("%s: Couldn't get xgmac sem\n", __func__);
  1211. return;
  1212. }
  1213. DUMP_XGMAC(qdev, PAUSE_SRC_LO);
  1214. DUMP_XGMAC(qdev, PAUSE_SRC_HI);
  1215. DUMP_XGMAC(qdev, GLOBAL_CFG);
  1216. DUMP_XGMAC(qdev, TX_CFG);
  1217. DUMP_XGMAC(qdev, RX_CFG);
  1218. DUMP_XGMAC(qdev, FLOW_CTL);
  1219. DUMP_XGMAC(qdev, PAUSE_OPCODE);
  1220. DUMP_XGMAC(qdev, PAUSE_TIMER);
  1221. DUMP_XGMAC(qdev, PAUSE_FRM_DEST_LO);
  1222. DUMP_XGMAC(qdev, PAUSE_FRM_DEST_HI);
  1223. DUMP_XGMAC(qdev, MAC_TX_PARAMS);
  1224. DUMP_XGMAC(qdev, MAC_RX_PARAMS);
  1225. DUMP_XGMAC(qdev, MAC_SYS_INT);
  1226. DUMP_XGMAC(qdev, MAC_SYS_INT_MASK);
  1227. DUMP_XGMAC(qdev, MAC_MGMT_INT);
  1228. DUMP_XGMAC(qdev, MAC_MGMT_IN_MASK);
  1229. DUMP_XGMAC(qdev, EXT_ARB_MODE);
  1230. ql_sem_unlock(qdev, qdev->xg_sem_mask);
  1231. }
  1232. static void ql_dump_ets_regs(struct ql_adapter *qdev)
  1233. {
  1234. }
  1235. static void ql_dump_cam_entries(struct ql_adapter *qdev)
  1236. {
  1237. int i;
  1238. u32 value[3];
  1239. i = ql_sem_spinlock(qdev, SEM_MAC_ADDR_MASK);
  1240. if (i)
  1241. return;
  1242. for (i = 0; i < 4; i++) {
  1243. if (ql_get_mac_addr_reg(qdev, MAC_ADDR_TYPE_CAM_MAC, i, value)) {
  1244. pr_err("%s: Failed read of mac index register\n",
  1245. __func__);
  1246. return;
  1247. } else {
  1248. if (value[0])
  1249. pr_err("%s: CAM index %d CAM Lookup Lower = 0x%.08x:%.08x, Output = 0x%.08x\n",
  1250. qdev->ndev->name, i, value[1], value[0],
  1251. value[2]);
  1252. }
  1253. }
  1254. for (i = 0; i < 32; i++) {
  1255. if (ql_get_mac_addr_reg
  1256. (qdev, MAC_ADDR_TYPE_MULTI_MAC, i, value)) {
  1257. pr_err("%s: Failed read of mac index register\n",
  1258. __func__);
  1259. return;
  1260. } else {
  1261. if (value[0])
  1262. pr_err("%s: MCAST index %d CAM Lookup Lower = 0x%.08x:%.08x\n",
  1263. qdev->ndev->name, i, value[1], value[0]);
  1264. }
  1265. }
  1266. ql_sem_unlock(qdev, SEM_MAC_ADDR_MASK);
  1267. }
  1268. void ql_dump_routing_entries(struct ql_adapter *qdev)
  1269. {
  1270. int i;
  1271. u32 value;
  1272. i = ql_sem_spinlock(qdev, SEM_RT_IDX_MASK);
  1273. if (i)
  1274. return;
  1275. for (i = 0; i < 16; i++) {
  1276. value = 0;
  1277. if (ql_get_routing_reg(qdev, i, &value)) {
  1278. pr_err("%s: Failed read of routing index register\n",
  1279. __func__);
  1280. return;
  1281. } else {
  1282. if (value)
  1283. pr_err("%s: Routing Mask %d = 0x%.08x\n",
  1284. qdev->ndev->name, i, value);
  1285. }
  1286. }
  1287. ql_sem_unlock(qdev, SEM_RT_IDX_MASK);
  1288. }
  1289. #define DUMP_REG(qdev, reg) \
  1290. pr_err("%-32s= 0x%x\n", #reg, ql_read32(qdev, reg))
  1291. void ql_dump_regs(struct ql_adapter *qdev)
  1292. {
  1293. pr_err("reg dump for function #%d\n", qdev->func);
  1294. DUMP_REG(qdev, SYS);
  1295. DUMP_REG(qdev, RST_FO);
  1296. DUMP_REG(qdev, FSC);
  1297. DUMP_REG(qdev, CSR);
  1298. DUMP_REG(qdev, ICB_RID);
  1299. DUMP_REG(qdev, ICB_L);
  1300. DUMP_REG(qdev, ICB_H);
  1301. DUMP_REG(qdev, CFG);
  1302. DUMP_REG(qdev, BIOS_ADDR);
  1303. DUMP_REG(qdev, STS);
  1304. DUMP_REG(qdev, INTR_EN);
  1305. DUMP_REG(qdev, INTR_MASK);
  1306. DUMP_REG(qdev, ISR1);
  1307. DUMP_REG(qdev, ISR2);
  1308. DUMP_REG(qdev, ISR3);
  1309. DUMP_REG(qdev, ISR4);
  1310. DUMP_REG(qdev, REV_ID);
  1311. DUMP_REG(qdev, FRC_ECC_ERR);
  1312. DUMP_REG(qdev, ERR_STS);
  1313. DUMP_REG(qdev, RAM_DBG_ADDR);
  1314. DUMP_REG(qdev, RAM_DBG_DATA);
  1315. DUMP_REG(qdev, ECC_ERR_CNT);
  1316. DUMP_REG(qdev, SEM);
  1317. DUMP_REG(qdev, GPIO_1);
  1318. DUMP_REG(qdev, GPIO_2);
  1319. DUMP_REG(qdev, GPIO_3);
  1320. DUMP_REG(qdev, XGMAC_ADDR);
  1321. DUMP_REG(qdev, XGMAC_DATA);
  1322. DUMP_REG(qdev, NIC_ETS);
  1323. DUMP_REG(qdev, CNA_ETS);
  1324. DUMP_REG(qdev, FLASH_ADDR);
  1325. DUMP_REG(qdev, FLASH_DATA);
  1326. DUMP_REG(qdev, CQ_STOP);
  1327. DUMP_REG(qdev, PAGE_TBL_RID);
  1328. DUMP_REG(qdev, WQ_PAGE_TBL_LO);
  1329. DUMP_REG(qdev, WQ_PAGE_TBL_HI);
  1330. DUMP_REG(qdev, CQ_PAGE_TBL_LO);
  1331. DUMP_REG(qdev, CQ_PAGE_TBL_HI);
  1332. DUMP_REG(qdev, COS_DFLT_CQ1);
  1333. DUMP_REG(qdev, COS_DFLT_CQ2);
  1334. DUMP_REG(qdev, SPLT_HDR);
  1335. DUMP_REG(qdev, FC_PAUSE_THRES);
  1336. DUMP_REG(qdev, NIC_PAUSE_THRES);
  1337. DUMP_REG(qdev, FC_ETHERTYPE);
  1338. DUMP_REG(qdev, FC_RCV_CFG);
  1339. DUMP_REG(qdev, NIC_RCV_CFG);
  1340. DUMP_REG(qdev, FC_COS_TAGS);
  1341. DUMP_REG(qdev, NIC_COS_TAGS);
  1342. DUMP_REG(qdev, MGMT_RCV_CFG);
  1343. DUMP_REG(qdev, XG_SERDES_ADDR);
  1344. DUMP_REG(qdev, XG_SERDES_DATA);
  1345. DUMP_REG(qdev, PRB_MX_ADDR);
  1346. DUMP_REG(qdev, PRB_MX_DATA);
  1347. ql_dump_intr_states(qdev);
  1348. ql_dump_xgmac_control_regs(qdev);
  1349. ql_dump_ets_regs(qdev);
  1350. ql_dump_cam_entries(qdev);
  1351. ql_dump_routing_entries(qdev);
  1352. }
  1353. #endif
  1354. #ifdef QL_STAT_DUMP
  1355. #define DUMP_STAT(qdev, stat) \
  1356. pr_err("%s = %ld\n", #stat, (unsigned long)qdev->nic_stats.stat)
  1357. void ql_dump_stat(struct ql_adapter *qdev)
  1358. {
  1359. pr_err("%s: Enter\n", __func__);
  1360. DUMP_STAT(qdev, tx_pkts);
  1361. DUMP_STAT(qdev, tx_bytes);
  1362. DUMP_STAT(qdev, tx_mcast_pkts);
  1363. DUMP_STAT(qdev, tx_bcast_pkts);
  1364. DUMP_STAT(qdev, tx_ucast_pkts);
  1365. DUMP_STAT(qdev, tx_ctl_pkts);
  1366. DUMP_STAT(qdev, tx_pause_pkts);
  1367. DUMP_STAT(qdev, tx_64_pkt);
  1368. DUMP_STAT(qdev, tx_65_to_127_pkt);
  1369. DUMP_STAT(qdev, tx_128_to_255_pkt);
  1370. DUMP_STAT(qdev, tx_256_511_pkt);
  1371. DUMP_STAT(qdev, tx_512_to_1023_pkt);
  1372. DUMP_STAT(qdev, tx_1024_to_1518_pkt);
  1373. DUMP_STAT(qdev, tx_1519_to_max_pkt);
  1374. DUMP_STAT(qdev, tx_undersize_pkt);
  1375. DUMP_STAT(qdev, tx_oversize_pkt);
  1376. DUMP_STAT(qdev, rx_bytes);
  1377. DUMP_STAT(qdev, rx_bytes_ok);
  1378. DUMP_STAT(qdev, rx_pkts);
  1379. DUMP_STAT(qdev, rx_pkts_ok);
  1380. DUMP_STAT(qdev, rx_bcast_pkts);
  1381. DUMP_STAT(qdev, rx_mcast_pkts);
  1382. DUMP_STAT(qdev, rx_ucast_pkts);
  1383. DUMP_STAT(qdev, rx_undersize_pkts);
  1384. DUMP_STAT(qdev, rx_oversize_pkts);
  1385. DUMP_STAT(qdev, rx_jabber_pkts);
  1386. DUMP_STAT(qdev, rx_undersize_fcerr_pkts);
  1387. DUMP_STAT(qdev, rx_drop_events);
  1388. DUMP_STAT(qdev, rx_fcerr_pkts);
  1389. DUMP_STAT(qdev, rx_align_err);
  1390. DUMP_STAT(qdev, rx_symbol_err);
  1391. DUMP_STAT(qdev, rx_mac_err);
  1392. DUMP_STAT(qdev, rx_ctl_pkts);
  1393. DUMP_STAT(qdev, rx_pause_pkts);
  1394. DUMP_STAT(qdev, rx_64_pkts);
  1395. DUMP_STAT(qdev, rx_65_to_127_pkts);
  1396. DUMP_STAT(qdev, rx_128_255_pkts);
  1397. DUMP_STAT(qdev, rx_256_511_pkts);
  1398. DUMP_STAT(qdev, rx_512_to_1023_pkts);
  1399. DUMP_STAT(qdev, rx_1024_to_1518_pkts);
  1400. DUMP_STAT(qdev, rx_1519_to_max_pkts);
  1401. DUMP_STAT(qdev, rx_len_err_pkts);
  1402. };
  1403. #endif
  1404. #ifdef QL_DEV_DUMP
  1405. #define DUMP_QDEV_FIELD(qdev, type, field) \
  1406. pr_err("qdev->%-24s = " type "\n", #field, qdev->field)
  1407. #define DUMP_QDEV_DMA_FIELD(qdev, field) \
  1408. pr_err("qdev->%-24s = %llx\n", #field, (unsigned long long)qdev->field)
  1409. #define DUMP_QDEV_ARRAY(qdev, type, array, index, field) \
  1410. pr_err("%s[%d].%s = " type "\n", \
  1411. #array, index, #field, qdev->array[index].field);
  1412. void ql_dump_qdev(struct ql_adapter *qdev)
  1413. {
  1414. int i;
  1415. DUMP_QDEV_FIELD(qdev, "%lx", flags);
  1416. DUMP_QDEV_FIELD(qdev, "%p", vlgrp);
  1417. DUMP_QDEV_FIELD(qdev, "%p", pdev);
  1418. DUMP_QDEV_FIELD(qdev, "%p", ndev);
  1419. DUMP_QDEV_FIELD(qdev, "%d", chip_rev_id);
  1420. DUMP_QDEV_FIELD(qdev, "%p", reg_base);
  1421. DUMP_QDEV_FIELD(qdev, "%p", doorbell_area);
  1422. DUMP_QDEV_FIELD(qdev, "%d", doorbell_area_size);
  1423. DUMP_QDEV_FIELD(qdev, "%x", msg_enable);
  1424. DUMP_QDEV_FIELD(qdev, "%p", rx_ring_shadow_reg_area);
  1425. DUMP_QDEV_DMA_FIELD(qdev, rx_ring_shadow_reg_dma);
  1426. DUMP_QDEV_FIELD(qdev, "%p", tx_ring_shadow_reg_area);
  1427. DUMP_QDEV_DMA_FIELD(qdev, tx_ring_shadow_reg_dma);
  1428. DUMP_QDEV_FIELD(qdev, "%d", intr_count);
  1429. if (qdev->msi_x_entry)
  1430. for (i = 0; i < qdev->intr_count; i++) {
  1431. DUMP_QDEV_ARRAY(qdev, "%d", msi_x_entry, i, vector);
  1432. DUMP_QDEV_ARRAY(qdev, "%d", msi_x_entry, i, entry);
  1433. }
  1434. for (i = 0; i < qdev->intr_count; i++) {
  1435. DUMP_QDEV_ARRAY(qdev, "%p", intr_context, i, qdev);
  1436. DUMP_QDEV_ARRAY(qdev, "%d", intr_context, i, intr);
  1437. DUMP_QDEV_ARRAY(qdev, "%d", intr_context, i, hooked);
  1438. DUMP_QDEV_ARRAY(qdev, "0x%08x", intr_context, i, intr_en_mask);
  1439. DUMP_QDEV_ARRAY(qdev, "0x%08x", intr_context, i, intr_dis_mask);
  1440. DUMP_QDEV_ARRAY(qdev, "0x%08x", intr_context, i, intr_read_mask);
  1441. }
  1442. DUMP_QDEV_FIELD(qdev, "%d", tx_ring_count);
  1443. DUMP_QDEV_FIELD(qdev, "%d", rx_ring_count);
  1444. DUMP_QDEV_FIELD(qdev, "%d", ring_mem_size);
  1445. DUMP_QDEV_FIELD(qdev, "%p", ring_mem);
  1446. DUMP_QDEV_FIELD(qdev, "%d", intr_count);
  1447. DUMP_QDEV_FIELD(qdev, "%p", tx_ring);
  1448. DUMP_QDEV_FIELD(qdev, "%d", rss_ring_count);
  1449. DUMP_QDEV_FIELD(qdev, "%p", rx_ring);
  1450. DUMP_QDEV_FIELD(qdev, "%d", default_rx_queue);
  1451. DUMP_QDEV_FIELD(qdev, "0x%08x", xg_sem_mask);
  1452. DUMP_QDEV_FIELD(qdev, "0x%08x", port_link_up);
  1453. DUMP_QDEV_FIELD(qdev, "0x%08x", port_init);
  1454. }
  1455. #endif
  1456. #ifdef QL_CB_DUMP
  1457. void ql_dump_wqicb(struct wqicb *wqicb)
  1458. {
  1459. pr_err("Dumping wqicb stuff...\n");
  1460. pr_err("wqicb->len = 0x%x\n", le16_to_cpu(wqicb->len));
  1461. pr_err("wqicb->flags = %x\n", le16_to_cpu(wqicb->flags));
  1462. pr_err("wqicb->cq_id_rss = %d\n",
  1463. le16_to_cpu(wqicb->cq_id_rss));
  1464. pr_err("wqicb->rid = 0x%x\n", le16_to_cpu(wqicb->rid));
  1465. pr_err("wqicb->wq_addr = 0x%llx\n",
  1466. (unsigned long long) le64_to_cpu(wqicb->addr));
  1467. pr_err("wqicb->wq_cnsmr_idx_addr = 0x%llx\n",
  1468. (unsigned long long) le64_to_cpu(wqicb->cnsmr_idx_addr));
  1469. }
  1470. void ql_dump_tx_ring(struct tx_ring *tx_ring)
  1471. {
  1472. if (tx_ring == NULL)
  1473. return;
  1474. pr_err("===================== Dumping tx_ring %d ===============\n",
  1475. tx_ring->wq_id);
  1476. pr_err("tx_ring->base = %p\n", tx_ring->wq_base);
  1477. pr_err("tx_ring->base_dma = 0x%llx\n",
  1478. (unsigned long long) tx_ring->wq_base_dma);
  1479. pr_err("tx_ring->cnsmr_idx_sh_reg, addr = 0x%p, value = %d\n",
  1480. tx_ring->cnsmr_idx_sh_reg,
  1481. tx_ring->cnsmr_idx_sh_reg
  1482. ? ql_read_sh_reg(tx_ring->cnsmr_idx_sh_reg) : 0);
  1483. pr_err("tx_ring->size = %d\n", tx_ring->wq_size);
  1484. pr_err("tx_ring->len = %d\n", tx_ring->wq_len);
  1485. pr_err("tx_ring->prod_idx_db_reg = %p\n", tx_ring->prod_idx_db_reg);
  1486. pr_err("tx_ring->valid_db_reg = %p\n", tx_ring->valid_db_reg);
  1487. pr_err("tx_ring->prod_idx = %d\n", tx_ring->prod_idx);
  1488. pr_err("tx_ring->cq_id = %d\n", tx_ring->cq_id);
  1489. pr_err("tx_ring->wq_id = %d\n", tx_ring->wq_id);
  1490. pr_err("tx_ring->q = %p\n", tx_ring->q);
  1491. pr_err("tx_ring->tx_count = %d\n", atomic_read(&tx_ring->tx_count));
  1492. }
  1493. void ql_dump_ricb(struct ricb *ricb)
  1494. {
  1495. int i;
  1496. pr_err("===================== Dumping ricb ===============\n");
  1497. pr_err("Dumping ricb stuff...\n");
  1498. pr_err("ricb->base_cq = %d\n", ricb->base_cq & 0x1f);
  1499. pr_err("ricb->flags = %s%s%s%s%s%s%s%s%s\n",
  1500. ricb->base_cq & RSS_L4K ? "RSS_L4K " : "",
  1501. ricb->flags & RSS_L6K ? "RSS_L6K " : "",
  1502. ricb->flags & RSS_LI ? "RSS_LI " : "",
  1503. ricb->flags & RSS_LB ? "RSS_LB " : "",
  1504. ricb->flags & RSS_LM ? "RSS_LM " : "",
  1505. ricb->flags & RSS_RI4 ? "RSS_RI4 " : "",
  1506. ricb->flags & RSS_RT4 ? "RSS_RT4 " : "",
  1507. ricb->flags & RSS_RI6 ? "RSS_RI6 " : "",
  1508. ricb->flags & RSS_RT6 ? "RSS_RT6 " : "");
  1509. pr_err("ricb->mask = 0x%.04x\n", le16_to_cpu(ricb->mask));
  1510. for (i = 0; i < 16; i++)
  1511. pr_err("ricb->hash_cq_id[%d] = 0x%.08x\n", i,
  1512. le32_to_cpu(ricb->hash_cq_id[i]));
  1513. for (i = 0; i < 10; i++)
  1514. pr_err("ricb->ipv6_hash_key[%d] = 0x%.08x\n", i,
  1515. le32_to_cpu(ricb->ipv6_hash_key[i]));
  1516. for (i = 0; i < 4; i++)
  1517. pr_err("ricb->ipv4_hash_key[%d] = 0x%.08x\n", i,
  1518. le32_to_cpu(ricb->ipv4_hash_key[i]));
  1519. }
  1520. void ql_dump_cqicb(struct cqicb *cqicb)
  1521. {
  1522. pr_err("Dumping cqicb stuff...\n");
  1523. pr_err("cqicb->msix_vect = %d\n", cqicb->msix_vect);
  1524. pr_err("cqicb->flags = %x\n", cqicb->flags);
  1525. pr_err("cqicb->len = %d\n", le16_to_cpu(cqicb->len));
  1526. pr_err("cqicb->addr = 0x%llx\n",
  1527. (unsigned long long) le64_to_cpu(cqicb->addr));
  1528. pr_err("cqicb->prod_idx_addr = 0x%llx\n",
  1529. (unsigned long long) le64_to_cpu(cqicb->prod_idx_addr));
  1530. pr_err("cqicb->pkt_delay = 0x%.04x\n",
  1531. le16_to_cpu(cqicb->pkt_delay));
  1532. pr_err("cqicb->irq_delay = 0x%.04x\n",
  1533. le16_to_cpu(cqicb->irq_delay));
  1534. pr_err("cqicb->lbq_addr = 0x%llx\n",
  1535. (unsigned long long) le64_to_cpu(cqicb->lbq_addr));
  1536. pr_err("cqicb->lbq_buf_size = 0x%.04x\n",
  1537. le16_to_cpu(cqicb->lbq_buf_size));
  1538. pr_err("cqicb->lbq_len = 0x%.04x\n",
  1539. le16_to_cpu(cqicb->lbq_len));
  1540. pr_err("cqicb->sbq_addr = 0x%llx\n",
  1541. (unsigned long long) le64_to_cpu(cqicb->sbq_addr));
  1542. pr_err("cqicb->sbq_buf_size = 0x%.04x\n",
  1543. le16_to_cpu(cqicb->sbq_buf_size));
  1544. pr_err("cqicb->sbq_len = 0x%.04x\n",
  1545. le16_to_cpu(cqicb->sbq_len));
  1546. }
  1547. void ql_dump_rx_ring(struct rx_ring *rx_ring)
  1548. {
  1549. if (rx_ring == NULL)
  1550. return;
  1551. pr_err("===================== Dumping rx_ring %d ===============\n",
  1552. rx_ring->cq_id);
  1553. pr_err("Dumping rx_ring %d, type = %s%s%s\n",
  1554. rx_ring->cq_id, rx_ring->type == DEFAULT_Q ? "DEFAULT" : "",
  1555. rx_ring->type == TX_Q ? "OUTBOUND COMPLETIONS" : "",
  1556. rx_ring->type == RX_Q ? "INBOUND_COMPLETIONS" : "");
  1557. pr_err("rx_ring->cqicb = %p\n", &rx_ring->cqicb);
  1558. pr_err("rx_ring->cq_base = %p\n", rx_ring->cq_base);
  1559. pr_err("rx_ring->cq_base_dma = %llx\n",
  1560. (unsigned long long) rx_ring->cq_base_dma);
  1561. pr_err("rx_ring->cq_size = %d\n", rx_ring->cq_size);
  1562. pr_err("rx_ring->cq_len = %d\n", rx_ring->cq_len);
  1563. pr_err("rx_ring->prod_idx_sh_reg, addr = 0x%p, value = %d\n",
  1564. rx_ring->prod_idx_sh_reg,
  1565. rx_ring->prod_idx_sh_reg
  1566. ? ql_read_sh_reg(rx_ring->prod_idx_sh_reg) : 0);
  1567. pr_err("rx_ring->prod_idx_sh_reg_dma = %llx\n",
  1568. (unsigned long long) rx_ring->prod_idx_sh_reg_dma);
  1569. pr_err("rx_ring->cnsmr_idx_db_reg = %p\n",
  1570. rx_ring->cnsmr_idx_db_reg);
  1571. pr_err("rx_ring->cnsmr_idx = %d\n", rx_ring->cnsmr_idx);
  1572. pr_err("rx_ring->curr_entry = %p\n", rx_ring->curr_entry);
  1573. pr_err("rx_ring->valid_db_reg = %p\n", rx_ring->valid_db_reg);
  1574. pr_err("rx_ring->lbq_base = %p\n", rx_ring->lbq_base);
  1575. pr_err("rx_ring->lbq_base_dma = %llx\n",
  1576. (unsigned long long) rx_ring->lbq_base_dma);
  1577. pr_err("rx_ring->lbq_base_indirect = %p\n",
  1578. rx_ring->lbq_base_indirect);
  1579. pr_err("rx_ring->lbq_base_indirect_dma = %llx\n",
  1580. (unsigned long long) rx_ring->lbq_base_indirect_dma);
  1581. pr_err("rx_ring->lbq = %p\n", rx_ring->lbq);
  1582. pr_err("rx_ring->lbq_len = %d\n", rx_ring->lbq_len);
  1583. pr_err("rx_ring->lbq_size = %d\n", rx_ring->lbq_size);
  1584. pr_err("rx_ring->lbq_prod_idx_db_reg = %p\n",
  1585. rx_ring->lbq_prod_idx_db_reg);
  1586. pr_err("rx_ring->lbq_prod_idx = %d\n", rx_ring->lbq_prod_idx);
  1587. pr_err("rx_ring->lbq_curr_idx = %d\n", rx_ring->lbq_curr_idx);
  1588. pr_err("rx_ring->lbq_clean_idx = %d\n", rx_ring->lbq_clean_idx);
  1589. pr_err("rx_ring->lbq_free_cnt = %d\n", rx_ring->lbq_free_cnt);
  1590. pr_err("rx_ring->lbq_buf_size = %d\n", rx_ring->lbq_buf_size);
  1591. pr_err("rx_ring->sbq_base = %p\n", rx_ring->sbq_base);
  1592. pr_err("rx_ring->sbq_base_dma = %llx\n",
  1593. (unsigned long long) rx_ring->sbq_base_dma);
  1594. pr_err("rx_ring->sbq_base_indirect = %p\n",
  1595. rx_ring->sbq_base_indirect);
  1596. pr_err("rx_ring->sbq_base_indirect_dma = %llx\n",
  1597. (unsigned long long) rx_ring->sbq_base_indirect_dma);
  1598. pr_err("rx_ring->sbq = %p\n", rx_ring->sbq);
  1599. pr_err("rx_ring->sbq_len = %d\n", rx_ring->sbq_len);
  1600. pr_err("rx_ring->sbq_size = %d\n", rx_ring->sbq_size);
  1601. pr_err("rx_ring->sbq_prod_idx_db_reg addr = %p\n",
  1602. rx_ring->sbq_prod_idx_db_reg);
  1603. pr_err("rx_ring->sbq_prod_idx = %d\n", rx_ring->sbq_prod_idx);
  1604. pr_err("rx_ring->sbq_curr_idx = %d\n", rx_ring->sbq_curr_idx);
  1605. pr_err("rx_ring->sbq_clean_idx = %d\n", rx_ring->sbq_clean_idx);
  1606. pr_err("rx_ring->sbq_free_cnt = %d\n", rx_ring->sbq_free_cnt);
  1607. pr_err("rx_ring->sbq_buf_size = %d\n", rx_ring->sbq_buf_size);
  1608. pr_err("rx_ring->cq_id = %d\n", rx_ring->cq_id);
  1609. pr_err("rx_ring->irq = %d\n", rx_ring->irq);
  1610. pr_err("rx_ring->cpu = %d\n", rx_ring->cpu);
  1611. pr_err("rx_ring->qdev = %p\n", rx_ring->qdev);
  1612. }
  1613. void ql_dump_hw_cb(struct ql_adapter *qdev, int size, u32 bit, u16 q_id)
  1614. {
  1615. void *ptr;
  1616. pr_err("%s: Enter\n", __func__);
  1617. ptr = kmalloc(size, GFP_ATOMIC);
  1618. if (ptr == NULL)
  1619. return;
  1620. if (ql_write_cfg(qdev, ptr, size, bit, q_id)) {
  1621. pr_err("%s: Failed to upload control block!\n", __func__);
  1622. goto fail_it;
  1623. }
  1624. switch (bit) {
  1625. case CFG_DRQ:
  1626. ql_dump_wqicb((struct wqicb *)ptr);
  1627. break;
  1628. case CFG_DCQ:
  1629. ql_dump_cqicb((struct cqicb *)ptr);
  1630. break;
  1631. case CFG_DR:
  1632. ql_dump_ricb((struct ricb *)ptr);
  1633. break;
  1634. default:
  1635. pr_err("%s: Invalid bit value = %x\n", __func__, bit);
  1636. break;
  1637. }
  1638. fail_it:
  1639. kfree(ptr);
  1640. }
  1641. #endif
  1642. #ifdef QL_OB_DUMP
  1643. void ql_dump_tx_desc(struct tx_buf_desc *tbd)
  1644. {
  1645. pr_err("tbd->addr = 0x%llx\n",
  1646. le64_to_cpu((u64) tbd->addr));
  1647. pr_err("tbd->len = %d\n",
  1648. le32_to_cpu(tbd->len & TX_DESC_LEN_MASK));
  1649. pr_err("tbd->flags = %s %s\n",
  1650. tbd->len & TX_DESC_C ? "C" : ".",
  1651. tbd->len & TX_DESC_E ? "E" : ".");
  1652. tbd++;
  1653. pr_err("tbd->addr = 0x%llx\n",
  1654. le64_to_cpu((u64) tbd->addr));
  1655. pr_err("tbd->len = %d\n",
  1656. le32_to_cpu(tbd->len & TX_DESC_LEN_MASK));
  1657. pr_err("tbd->flags = %s %s\n",
  1658. tbd->len & TX_DESC_C ? "C" : ".",
  1659. tbd->len & TX_DESC_E ? "E" : ".");
  1660. tbd++;
  1661. pr_err("tbd->addr = 0x%llx\n",
  1662. le64_to_cpu((u64) tbd->addr));
  1663. pr_err("tbd->len = %d\n",
  1664. le32_to_cpu(tbd->len & TX_DESC_LEN_MASK));
  1665. pr_err("tbd->flags = %s %s\n",
  1666. tbd->len & TX_DESC_C ? "C" : ".",
  1667. tbd->len & TX_DESC_E ? "E" : ".");
  1668. }
  1669. void ql_dump_ob_mac_iocb(struct ob_mac_iocb_req *ob_mac_iocb)
  1670. {
  1671. struct ob_mac_tso_iocb_req *ob_mac_tso_iocb =
  1672. (struct ob_mac_tso_iocb_req *)ob_mac_iocb;
  1673. struct tx_buf_desc *tbd;
  1674. u16 frame_len;
  1675. pr_err("%s\n", __func__);
  1676. pr_err("opcode = %s\n",
  1677. (ob_mac_iocb->opcode == OPCODE_OB_MAC_IOCB) ? "MAC" : "TSO");
  1678. pr_err("flags1 = %s %s %s %s %s\n",
  1679. ob_mac_tso_iocb->flags1 & OB_MAC_TSO_IOCB_OI ? "OI" : "",
  1680. ob_mac_tso_iocb->flags1 & OB_MAC_TSO_IOCB_I ? "I" : "",
  1681. ob_mac_tso_iocb->flags1 & OB_MAC_TSO_IOCB_D ? "D" : "",
  1682. ob_mac_tso_iocb->flags1 & OB_MAC_TSO_IOCB_IP4 ? "IP4" : "",
  1683. ob_mac_tso_iocb->flags1 & OB_MAC_TSO_IOCB_IP6 ? "IP6" : "");
  1684. pr_err("flags2 = %s %s %s\n",
  1685. ob_mac_tso_iocb->flags2 & OB_MAC_TSO_IOCB_LSO ? "LSO" : "",
  1686. ob_mac_tso_iocb->flags2 & OB_MAC_TSO_IOCB_UC ? "UC" : "",
  1687. ob_mac_tso_iocb->flags2 & OB_MAC_TSO_IOCB_TC ? "TC" : "");
  1688. pr_err("flags3 = %s %s %s\n",
  1689. ob_mac_tso_iocb->flags3 & OB_MAC_TSO_IOCB_IC ? "IC" : "",
  1690. ob_mac_tso_iocb->flags3 & OB_MAC_TSO_IOCB_DFP ? "DFP" : "",
  1691. ob_mac_tso_iocb->flags3 & OB_MAC_TSO_IOCB_V ? "V" : "");
  1692. pr_err("tid = %x\n", ob_mac_iocb->tid);
  1693. pr_err("txq_idx = %d\n", ob_mac_iocb->txq_idx);
  1694. pr_err("vlan_tci = %x\n", ob_mac_tso_iocb->vlan_tci);
  1695. if (ob_mac_iocb->opcode == OPCODE_OB_MAC_TSO_IOCB) {
  1696. pr_err("frame_len = %d\n",
  1697. le32_to_cpu(ob_mac_tso_iocb->frame_len));
  1698. pr_err("mss = %d\n",
  1699. le16_to_cpu(ob_mac_tso_iocb->mss));
  1700. pr_err("prot_hdr_len = %d\n",
  1701. le16_to_cpu(ob_mac_tso_iocb->total_hdrs_len));
  1702. pr_err("hdr_offset = 0x%.04x\n",
  1703. le16_to_cpu(ob_mac_tso_iocb->net_trans_offset));
  1704. frame_len = le32_to_cpu(ob_mac_tso_iocb->frame_len);
  1705. } else {
  1706. pr_err("frame_len = %d\n",
  1707. le16_to_cpu(ob_mac_iocb->frame_len));
  1708. frame_len = le16_to_cpu(ob_mac_iocb->frame_len);
  1709. }
  1710. tbd = &ob_mac_iocb->tbd[0];
  1711. ql_dump_tx_desc(tbd);
  1712. }
  1713. void ql_dump_ob_mac_rsp(struct ob_mac_iocb_rsp *ob_mac_rsp)
  1714. {
  1715. pr_err("%s\n", __func__);
  1716. pr_err("opcode = %d\n", ob_mac_rsp->opcode);
  1717. pr_err("flags = %s %s %s %s %s %s %s\n",
  1718. ob_mac_rsp->flags1 & OB_MAC_IOCB_RSP_OI ? "OI" : ".",
  1719. ob_mac_rsp->flags1 & OB_MAC_IOCB_RSP_I ? "I" : ".",
  1720. ob_mac_rsp->flags1 & OB_MAC_IOCB_RSP_E ? "E" : ".",
  1721. ob_mac_rsp->flags1 & OB_MAC_IOCB_RSP_S ? "S" : ".",
  1722. ob_mac_rsp->flags1 & OB_MAC_IOCB_RSP_L ? "L" : ".",
  1723. ob_mac_rsp->flags1 & OB_MAC_IOCB_RSP_P ? "P" : ".",
  1724. ob_mac_rsp->flags2 & OB_MAC_IOCB_RSP_B ? "B" : ".");
  1725. pr_err("tid = %x\n", ob_mac_rsp->tid);
  1726. }
  1727. #endif
  1728. #ifdef QL_IB_DUMP
  1729. void ql_dump_ib_mac_rsp(struct ib_mac_iocb_rsp *ib_mac_rsp)
  1730. {
  1731. pr_err("%s\n", __func__);
  1732. pr_err("opcode = 0x%x\n", ib_mac_rsp->opcode);
  1733. pr_err("flags1 = %s%s%s%s%s%s\n",
  1734. ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_OI ? "OI " : "",
  1735. ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_I ? "I " : "",
  1736. ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_TE ? "TE " : "",
  1737. ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_NU ? "NU " : "",
  1738. ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_IE ? "IE " : "",
  1739. ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_B ? "B " : "");
  1740. if (ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_M_MASK)
  1741. pr_err("%s%s%s Multicast\n",
  1742. (ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_M_MASK) ==
  1743. IB_MAC_IOCB_RSP_M_HASH ? "Hash" : "",
  1744. (ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_M_MASK) ==
  1745. IB_MAC_IOCB_RSP_M_REG ? "Registered" : "",
  1746. (ib_mac_rsp->flags1 & IB_MAC_IOCB_RSP_M_MASK) ==
  1747. IB_MAC_IOCB_RSP_M_PROM ? "Promiscuous" : "");
  1748. pr_err("flags2 = %s%s%s%s%s\n",
  1749. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_P) ? "P " : "",
  1750. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_V) ? "V " : "",
  1751. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_U) ? "U " : "",
  1752. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_T) ? "T " : "",
  1753. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_FO) ? "FO " : "");
  1754. if (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_ERR_MASK)
  1755. pr_err("%s%s%s%s%s error\n",
  1756. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_ERR_MASK) ==
  1757. IB_MAC_IOCB_RSP_ERR_OVERSIZE ? "oversize" : "",
  1758. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_ERR_MASK) ==
  1759. IB_MAC_IOCB_RSP_ERR_UNDERSIZE ? "undersize" : "",
  1760. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_ERR_MASK) ==
  1761. IB_MAC_IOCB_RSP_ERR_PREAMBLE ? "preamble" : "",
  1762. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_ERR_MASK) ==
  1763. IB_MAC_IOCB_RSP_ERR_FRAME_LEN ? "frame length" : "",
  1764. (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_ERR_MASK) ==
  1765. IB_MAC_IOCB_RSP_ERR_CRC ? "CRC" : "");
  1766. pr_err("flags3 = %s%s\n",
  1767. ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_DS ? "DS " : "",
  1768. ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_DL ? "DL " : "");
  1769. if (ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_RSS_MASK)
  1770. pr_err("RSS flags = %s%s%s%s\n",
  1771. ((ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_RSS_MASK) ==
  1772. IB_MAC_IOCB_RSP_M_IPV4) ? "IPv4 RSS" : "",
  1773. ((ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_RSS_MASK) ==
  1774. IB_MAC_IOCB_RSP_M_IPV6) ? "IPv6 RSS " : "",
  1775. ((ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_RSS_MASK) ==
  1776. IB_MAC_IOCB_RSP_M_TCP_V4) ? "TCP/IPv4 RSS" : "",
  1777. ((ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_RSS_MASK) ==
  1778. IB_MAC_IOCB_RSP_M_TCP_V6) ? "TCP/IPv6 RSS" : "");
  1779. pr_err("data_len = %d\n",
  1780. le32_to_cpu(ib_mac_rsp->data_len));
  1781. pr_err("data_addr = 0x%llx\n",
  1782. (unsigned long long) le64_to_cpu(ib_mac_rsp->data_addr));
  1783. if (ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_RSS_MASK)
  1784. pr_err("rss = %x\n",
  1785. le32_to_cpu(ib_mac_rsp->rss));
  1786. if (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_V)
  1787. pr_err("vlan_id = %x\n",
  1788. le16_to_cpu(ib_mac_rsp->vlan_id));
  1789. pr_err("flags4 = %s%s%s\n",
  1790. ib_mac_rsp->flags4 & IB_MAC_IOCB_RSP_HV ? "HV " : "",
  1791. ib_mac_rsp->flags4 & IB_MAC_IOCB_RSP_HS ? "HS " : "",
  1792. ib_mac_rsp->flags4 & IB_MAC_IOCB_RSP_HL ? "HL " : "");
  1793. if (ib_mac_rsp->flags4 & IB_MAC_IOCB_RSP_HV) {
  1794. pr_err("hdr length = %d\n",
  1795. le32_to_cpu(ib_mac_rsp->hdr_len));
  1796. pr_err("hdr addr = 0x%llx\n",
  1797. (unsigned long long) le64_to_cpu(ib_mac_rsp->hdr_addr));
  1798. }
  1799. }
  1800. #endif
  1801. #ifdef QL_ALL_DUMP
  1802. void ql_dump_all(struct ql_adapter *qdev)
  1803. {
  1804. int i;
  1805. QL_DUMP_REGS(qdev);
  1806. QL_DUMP_QDEV(qdev);
  1807. for (i = 0; i < qdev->tx_ring_count; i++) {
  1808. QL_DUMP_TX_RING(&qdev->tx_ring[i]);
  1809. QL_DUMP_WQICB((struct wqicb *)&qdev->tx_ring[i]);
  1810. }
  1811. for (i = 0; i < qdev->rx_ring_count; i++) {
  1812. QL_DUMP_RX_RING(&qdev->rx_ring[i]);
  1813. QL_DUMP_CQICB((struct cqicb *)&qdev->rx_ring[i]);
  1814. }
  1815. }
  1816. #endif