w83977af_ir.c 30 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285
  1. /*********************************************************************
  2. *
  3. * Filename: w83977af_ir.c
  4. * Version: 1.0
  5. * Description: FIR driver for the Winbond W83977AF Super I/O chip
  6. * Status: Experimental.
  7. * Author: Paul VanderSpek
  8. * Created at: Wed Nov 4 11:46:16 1998
  9. * Modified at: Fri Jan 28 12:10:59 2000
  10. * Modified by: Dag Brattli <dagb@cs.uit.no>
  11. *
  12. * Copyright (c) 1998-2000 Dag Brattli <dagb@cs.uit.no>
  13. * Copyright (c) 1998-1999 Rebel.com
  14. *
  15. * This program is free software; you can redistribute it and/or
  16. * modify it under the terms of the GNU General Public License as
  17. * published by the Free Software Foundation; either version 2 of
  18. * the License, or (at your option) any later version.
  19. *
  20. * Neither Paul VanderSpek nor Rebel.com admit liability nor provide
  21. * warranty for any of this software. This material is provided "AS-IS"
  22. * and at no charge.
  23. *
  24. * If you find bugs in this file, its very likely that the same bug
  25. * will also be in pc87108.c since the implementations are quite
  26. * similar.
  27. *
  28. * Notice that all functions that needs to access the chip in _any_
  29. * way, must save BSR register on entry, and restore it on exit.
  30. * It is _very_ important to follow this policy!
  31. *
  32. * __u8 bank;
  33. *
  34. * bank = inb( iobase+BSR);
  35. *
  36. * do_your_stuff_here();
  37. *
  38. * outb( bank, iobase+BSR);
  39. *
  40. ********************************************************************/
  41. #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
  42. #include <linux/module.h>
  43. #include <linux/kernel.h>
  44. #include <linux/types.h>
  45. #include <linux/skbuff.h>
  46. #include <linux/netdevice.h>
  47. #include <linux/ioport.h>
  48. #include <linux/delay.h>
  49. #include <linux/init.h>
  50. #include <linux/interrupt.h>
  51. #include <linux/rtnetlink.h>
  52. #include <linux/dma-mapping.h>
  53. #include <linux/gfp.h>
  54. #include <asm/io.h>
  55. #include <asm/dma.h>
  56. #include <asm/byteorder.h>
  57. #include <net/irda/irda.h>
  58. #include <net/irda/wrapper.h>
  59. #include <net/irda/irda_device.h>
  60. #include "w83977af.h"
  61. #include "w83977af_ir.h"
  62. #define CONFIG_USE_W977_PNP /* Currently needed */
  63. #define PIO_MAX_SPEED 115200
  64. static char *driver_name = "w83977af_ir";
  65. static int qos_mtt_bits = 0x07; /* 1 ms or more */
  66. #define CHIP_IO_EXTENT 8
  67. static unsigned int io[] = { 0x180, ~0, ~0, ~0 };
  68. #ifdef CONFIG_ARCH_NETWINDER /* Adjust to NetWinder differences */
  69. static unsigned int irq[] = { 6, 0, 0, 0 };
  70. #else
  71. static unsigned int irq[] = { 11, 0, 0, 0 };
  72. #endif
  73. static unsigned int dma[] = { 1, 0, 0, 0 };
  74. static unsigned int efbase[] = { W977_EFIO_BASE, W977_EFIO2_BASE };
  75. static unsigned int efio = W977_EFIO_BASE;
  76. static struct w83977af_ir *dev_self[] = { NULL, NULL, NULL, NULL};
  77. /* Some prototypes */
  78. static int w83977af_open(int i, unsigned int iobase, unsigned int irq,
  79. unsigned int dma);
  80. static int w83977af_close(struct w83977af_ir *self);
  81. static int w83977af_probe(int iobase, int irq, int dma);
  82. static int w83977af_dma_receive(struct w83977af_ir *self);
  83. static int w83977af_dma_receive_complete(struct w83977af_ir *self);
  84. static netdev_tx_t w83977af_hard_xmit(struct sk_buff *skb,
  85. struct net_device *dev);
  86. static int w83977af_pio_write(int iobase, __u8 *buf, int len, int fifo_size);
  87. static void w83977af_dma_write(struct w83977af_ir *self, int iobase);
  88. static void w83977af_change_speed(struct w83977af_ir *self, __u32 speed);
  89. static int w83977af_is_receiving(struct w83977af_ir *self);
  90. static int w83977af_net_open(struct net_device *dev);
  91. static int w83977af_net_close(struct net_device *dev);
  92. static int w83977af_net_ioctl(struct net_device *dev, struct ifreq *rq, int cmd);
  93. /*
  94. * Function w83977af_init ()
  95. *
  96. * Initialize chip. Just try to find out how many chips we are dealing with
  97. * and where they are
  98. */
  99. static int __init w83977af_init(void)
  100. {
  101. int i;
  102. for (i = 0; i < ARRAY_SIZE(dev_self) && io[i] < 2000; i++) {
  103. if (w83977af_open(i, io[i], irq[i], dma[i]) == 0)
  104. return 0;
  105. }
  106. return -ENODEV;
  107. }
  108. /*
  109. * Function w83977af_cleanup ()
  110. *
  111. * Close all configured chips
  112. *
  113. */
  114. static void __exit w83977af_cleanup(void)
  115. {
  116. int i;
  117. for (i = 0; i < ARRAY_SIZE(dev_self); i++) {
  118. if (dev_self[i])
  119. w83977af_close(dev_self[i]);
  120. }
  121. }
  122. static const struct net_device_ops w83977_netdev_ops = {
  123. .ndo_open = w83977af_net_open,
  124. .ndo_stop = w83977af_net_close,
  125. .ndo_start_xmit = w83977af_hard_xmit,
  126. .ndo_do_ioctl = w83977af_net_ioctl,
  127. };
  128. /*
  129. * Function w83977af_open (iobase, irq)
  130. *
  131. * Open driver instance
  132. *
  133. */
  134. static int w83977af_open(int i, unsigned int iobase, unsigned int irq,
  135. unsigned int dma)
  136. {
  137. struct net_device *dev;
  138. struct w83977af_ir *self;
  139. int err;
  140. /* Lock the port that we need */
  141. if (!request_region(iobase, CHIP_IO_EXTENT, driver_name)) {
  142. pr_debug("%s: can't get iobase of 0x%03x\n",
  143. __func__, iobase);
  144. return -ENODEV;
  145. }
  146. if (w83977af_probe(iobase, irq, dma) == -1) {
  147. err = -1;
  148. goto err_out;
  149. }
  150. /*
  151. * Allocate new instance of the driver
  152. */
  153. dev = alloc_irdadev(sizeof(struct w83977af_ir));
  154. if (!dev) {
  155. pr_err("IrDA: Can't allocate memory for IrDA control block!\n");
  156. err = -ENOMEM;
  157. goto err_out;
  158. }
  159. self = netdev_priv(dev);
  160. spin_lock_init(&self->lock);
  161. /* Initialize IO */
  162. self->io.fir_base = iobase;
  163. self->io.irq = irq;
  164. self->io.fir_ext = CHIP_IO_EXTENT;
  165. self->io.dma = dma;
  166. self->io.fifo_size = 32;
  167. /* Initialize QoS for this device */
  168. irda_init_max_qos_capabilies(&self->qos);
  169. /* The only value we must override it the baudrate */
  170. /* FIXME: The HP HDLS-1100 does not support 1152000! */
  171. self->qos.baud_rate.bits = IR_9600 | IR_19200 | IR_38400 | IR_57600 |
  172. IR_115200 | IR_576000 | IR_1152000 | (IR_4000000 << 8);
  173. /* The HP HDLS-1100 needs 1 ms according to the specs */
  174. self->qos.min_turn_time.bits = qos_mtt_bits;
  175. irda_qos_bits_to_value(&self->qos);
  176. /* Max DMA buffer size needed = (data_size + 6) * (window_size) + 6; */
  177. self->rx_buff.truesize = 14384;
  178. self->tx_buff.truesize = 4000;
  179. /* Allocate memory if needed */
  180. self->rx_buff.head =
  181. dma_zalloc_coherent(NULL, self->rx_buff.truesize,
  182. &self->rx_buff_dma, GFP_KERNEL);
  183. if (!self->rx_buff.head) {
  184. err = -ENOMEM;
  185. goto err_out1;
  186. }
  187. self->tx_buff.head =
  188. dma_zalloc_coherent(NULL, self->tx_buff.truesize,
  189. &self->tx_buff_dma, GFP_KERNEL);
  190. if (!self->tx_buff.head) {
  191. err = -ENOMEM;
  192. goto err_out2;
  193. }
  194. self->rx_buff.in_frame = FALSE;
  195. self->rx_buff.state = OUTSIDE_FRAME;
  196. self->tx_buff.data = self->tx_buff.head;
  197. self->rx_buff.data = self->rx_buff.head;
  198. self->netdev = dev;
  199. dev->netdev_ops = &w83977_netdev_ops;
  200. err = register_netdev(dev);
  201. if (err) {
  202. net_err_ratelimited("%s:, register_netdevice() failed!\n",
  203. __func__);
  204. goto err_out3;
  205. }
  206. net_info_ratelimited("IrDA: Registered device %s\n", dev->name);
  207. /* Need to store self somewhere */
  208. dev_self[i] = self;
  209. return 0;
  210. err_out3:
  211. dma_free_coherent(NULL, self->tx_buff.truesize,
  212. self->tx_buff.head, self->tx_buff_dma);
  213. err_out2:
  214. dma_free_coherent(NULL, self->rx_buff.truesize,
  215. self->rx_buff.head, self->rx_buff_dma);
  216. err_out1:
  217. free_netdev(dev);
  218. err_out:
  219. release_region(iobase, CHIP_IO_EXTENT);
  220. return err;
  221. }
  222. /*
  223. * Function w83977af_close (self)
  224. *
  225. * Close driver instance
  226. *
  227. */
  228. static int w83977af_close(struct w83977af_ir *self)
  229. {
  230. int iobase;
  231. iobase = self->io.fir_base;
  232. #ifdef CONFIG_USE_W977_PNP
  233. /* enter PnP configuration mode */
  234. w977_efm_enter(efio);
  235. w977_select_device(W977_DEVICE_IR, efio);
  236. /* Deactivate device */
  237. w977_write_reg(0x30, 0x00, efio);
  238. w977_efm_exit(efio);
  239. #endif /* CONFIG_USE_W977_PNP */
  240. /* Remove netdevice */
  241. unregister_netdev(self->netdev);
  242. /* Release the PORT that this driver is using */
  243. pr_debug("%s: Releasing Region %03x\n", __func__, self->io.fir_base);
  244. release_region(self->io.fir_base, self->io.fir_ext);
  245. if (self->tx_buff.head)
  246. dma_free_coherent(NULL, self->tx_buff.truesize,
  247. self->tx_buff.head, self->tx_buff_dma);
  248. if (self->rx_buff.head)
  249. dma_free_coherent(NULL, self->rx_buff.truesize,
  250. self->rx_buff.head, self->rx_buff_dma);
  251. free_netdev(self->netdev);
  252. return 0;
  253. }
  254. static int w83977af_probe(int iobase, int irq, int dma)
  255. {
  256. int version;
  257. int i;
  258. for (i = 0; i < 2; i++) {
  259. #ifdef CONFIG_USE_W977_PNP
  260. /* Enter PnP configuration mode */
  261. w977_efm_enter(efbase[i]);
  262. w977_select_device(W977_DEVICE_IR, efbase[i]);
  263. /* Configure PnP port, IRQ, and DMA channel */
  264. w977_write_reg(0x60, (iobase >> 8) & 0xff, efbase[i]);
  265. w977_write_reg(0x61, (iobase) & 0xff, efbase[i]);
  266. w977_write_reg(0x70, irq, efbase[i]);
  267. #ifdef CONFIG_ARCH_NETWINDER
  268. /* Netwinder uses 1 higher than Linux */
  269. w977_write_reg(0x74, dma + 1, efbase[i]);
  270. #else
  271. w977_write_reg(0x74, dma, efbase[i]);
  272. #endif /* CONFIG_ARCH_NETWINDER */
  273. w977_write_reg(0x75, 0x04, efbase[i]);/* Disable Tx DMA */
  274. /* Set append hardware CRC, enable IR bank selection */
  275. w977_write_reg(0xf0, APEDCRC | ENBNKSEL, efbase[i]);
  276. /* Activate device */
  277. w977_write_reg(0x30, 0x01, efbase[i]);
  278. w977_efm_exit(efbase[i]);
  279. #endif /* CONFIG_USE_W977_PNP */
  280. /* Disable Advanced mode */
  281. switch_bank(iobase, SET2);
  282. outb(iobase + 2, 0x00);
  283. /* Turn on UART (global) interrupts */
  284. switch_bank(iobase, SET0);
  285. outb(HCR_EN_IRQ, iobase + HCR);
  286. /* Switch to advanced mode */
  287. switch_bank(iobase, SET2);
  288. outb(inb(iobase + ADCR1) | ADCR1_ADV_SL, iobase + ADCR1);
  289. /* Set default IR-mode */
  290. switch_bank(iobase, SET0);
  291. outb(HCR_SIR, iobase + HCR);
  292. /* Read the Advanced IR ID */
  293. switch_bank(iobase, SET3);
  294. version = inb(iobase + AUID);
  295. /* Should be 0x1? */
  296. if (0x10 == (version & 0xf0)) {
  297. efio = efbase[i];
  298. /* Set FIFO size to 32 */
  299. switch_bank(iobase, SET2);
  300. outb(ADCR2_RXFS32 | ADCR2_TXFS32, iobase + ADCR2);
  301. /* Set FIFO threshold to TX17, RX16 */
  302. switch_bank(iobase, SET0);
  303. outb(UFR_RXTL | UFR_TXTL | UFR_TXF_RST | UFR_RXF_RST |
  304. UFR_EN_FIFO, iobase + UFR);
  305. /* Receiver frame length */
  306. switch_bank(iobase, SET4);
  307. outb(2048 & 0xff, iobase + 6);
  308. outb((2048 >> 8) & 0x1f, iobase + 7);
  309. /*
  310. * Init HP HSDL-1100 transceiver.
  311. *
  312. * Set IRX_MSL since we have 2 * receive paths IRRX,
  313. * and IRRXH. Clear IRSL0D since we want IRSL0 * to
  314. * be a input pin used for IRRXH
  315. *
  316. * IRRX pin 37 connected to receiver
  317. * IRTX pin 38 connected to transmitter
  318. * FIRRX pin 39 connected to receiver (IRSL0)
  319. * CIRRX pin 40 connected to pin 37
  320. */
  321. switch_bank(iobase, SET7);
  322. outb(0x40, iobase + 7);
  323. net_info_ratelimited("W83977AF (IR) driver loaded. Version: 0x%02x\n",
  324. version);
  325. return 0;
  326. } else {
  327. /* Try next extented function register address */
  328. pr_debug("%s: Wrong chip version\n", __func__);
  329. }
  330. }
  331. return -1;
  332. }
  333. static void w83977af_change_speed(struct w83977af_ir *self, __u32 speed)
  334. {
  335. int ir_mode = HCR_SIR;
  336. int iobase;
  337. __u8 set;
  338. iobase = self->io.fir_base;
  339. /* Update accounting for new speed */
  340. self->io.speed = speed;
  341. /* Save current bank */
  342. set = inb(iobase + SSR);
  343. /* Disable interrupts */
  344. switch_bank(iobase, SET0);
  345. outb(0, iobase + ICR);
  346. /* Select Set 2 */
  347. switch_bank(iobase, SET2);
  348. outb(0x00, iobase + ABHL);
  349. switch (speed) {
  350. case 9600: outb(0x0c, iobase + ABLL); break;
  351. case 19200: outb(0x06, iobase + ABLL); break;
  352. case 38400: outb(0x03, iobase + ABLL); break;
  353. case 57600: outb(0x02, iobase + ABLL); break;
  354. case 115200: outb(0x01, iobase + ABLL); break;
  355. case 576000:
  356. ir_mode = HCR_MIR_576;
  357. pr_debug("%s: handling baud of 576000\n", __func__);
  358. break;
  359. case 1152000:
  360. ir_mode = HCR_MIR_1152;
  361. pr_debug("%s: handling baud of 1152000\n", __func__);
  362. break;
  363. case 4000000:
  364. ir_mode = HCR_FIR;
  365. pr_debug("%s: handling baud of 4000000\n", __func__);
  366. break;
  367. default:
  368. ir_mode = HCR_FIR;
  369. pr_debug("%s: unknown baud rate of %d\n", __func__, speed);
  370. break;
  371. }
  372. /* Set speed mode */
  373. switch_bank(iobase, SET0);
  374. outb(ir_mode, iobase + HCR);
  375. /* set FIFO size to 32 */
  376. switch_bank(iobase, SET2);
  377. outb(ADCR2_RXFS32 | ADCR2_TXFS32, iobase + ADCR2);
  378. /* set FIFO threshold to TX17, RX16 */
  379. switch_bank(iobase, SET0);
  380. outb(0x00, iobase + UFR); /* Reset */
  381. outb(UFR_EN_FIFO, iobase + UFR); /* First we must enable FIFO */
  382. outb(0xa7, iobase + UFR);
  383. netif_wake_queue(self->netdev);
  384. /* Enable some interrupts so we can receive frames */
  385. switch_bank(iobase, SET0);
  386. if (speed > PIO_MAX_SPEED) {
  387. outb(ICR_EFSFI, iobase + ICR);
  388. w83977af_dma_receive(self);
  389. } else {
  390. outb(ICR_ERBRI, iobase + ICR);
  391. }
  392. /* Restore SSR */
  393. outb(set, iobase + SSR);
  394. }
  395. /*
  396. * Function w83977af_hard_xmit (skb, dev)
  397. *
  398. * Sets up a DMA transfer to send the current frame.
  399. *
  400. */
  401. static netdev_tx_t w83977af_hard_xmit(struct sk_buff *skb,
  402. struct net_device *dev)
  403. {
  404. struct w83977af_ir *self;
  405. __s32 speed;
  406. int iobase;
  407. __u8 set;
  408. int mtt;
  409. self = netdev_priv(dev);
  410. iobase = self->io.fir_base;
  411. pr_debug("%s: %ld, skb->len=%d\n", __func__, jiffies, (int)skb->len);
  412. /* Lock transmit buffer */
  413. netif_stop_queue(dev);
  414. /* Check if we need to change the speed */
  415. speed = irda_get_next_speed(skb);
  416. if ((speed != self->io.speed) && (speed != -1)) {
  417. /* Check for empty frame */
  418. if (!skb->len) {
  419. w83977af_change_speed(self, speed);
  420. dev_kfree_skb(skb);
  421. return NETDEV_TX_OK;
  422. }
  423. self->new_speed = speed;
  424. }
  425. /* Save current set */
  426. set = inb(iobase + SSR);
  427. /* Decide if we should use PIO or DMA transfer */
  428. if (self->io.speed > PIO_MAX_SPEED) {
  429. self->tx_buff.data = self->tx_buff.head;
  430. skb_copy_from_linear_data(skb, self->tx_buff.data, skb->len);
  431. self->tx_buff.len = skb->len;
  432. mtt = irda_get_mtt(skb);
  433. pr_debug("%s: %ld, mtt=%d\n", __func__, jiffies, mtt);
  434. if (mtt > 1000)
  435. mdelay(mtt / 1000);
  436. else if (mtt)
  437. udelay(mtt);
  438. /* Enable DMA interrupt */
  439. switch_bank(iobase, SET0);
  440. outb(ICR_EDMAI, iobase + ICR);
  441. w83977af_dma_write(self, iobase);
  442. } else {
  443. self->tx_buff.data = self->tx_buff.head;
  444. self->tx_buff.len = async_wrap_skb(skb, self->tx_buff.data,
  445. self->tx_buff.truesize);
  446. /* Add interrupt on tx low level (will fire immediately) */
  447. switch_bank(iobase, SET0);
  448. outb(ICR_ETXTHI, iobase + ICR);
  449. }
  450. dev_kfree_skb(skb);
  451. /* Restore set register */
  452. outb(set, iobase + SSR);
  453. return NETDEV_TX_OK;
  454. }
  455. /*
  456. * Function w83977af_dma_write (self, iobase)
  457. *
  458. * Send frame using DMA
  459. *
  460. */
  461. static void w83977af_dma_write(struct w83977af_ir *self, int iobase)
  462. {
  463. __u8 set;
  464. pr_debug("%s: len=%d\n", __func__, self->tx_buff.len);
  465. /* Save current set */
  466. set = inb(iobase + SSR);
  467. /* Disable DMA */
  468. switch_bank(iobase, SET0);
  469. outb(inb(iobase + HCR) & ~HCR_EN_DMA, iobase + HCR);
  470. /* Choose transmit DMA channel */
  471. switch_bank(iobase, SET2);
  472. outb(ADCR1_D_CHSW | /*ADCR1_DMA_F|*/ADCR1_ADV_SL, iobase + ADCR1);
  473. irda_setup_dma(self->io.dma, self->tx_buff_dma, self->tx_buff.len,
  474. DMA_MODE_WRITE);
  475. self->io.direction = IO_XMIT;
  476. /* Enable DMA */
  477. switch_bank(iobase, SET0);
  478. outb(inb(iobase + HCR) | HCR_EN_DMA | HCR_TX_WT, iobase + HCR);
  479. /* Restore set register */
  480. outb(set, iobase + SSR);
  481. }
  482. /*
  483. * Function w83977af_pio_write (iobase, buf, len, fifo_size)
  484. *
  485. *
  486. *
  487. */
  488. static int w83977af_pio_write(int iobase, __u8 *buf, int len, int fifo_size)
  489. {
  490. int actual = 0;
  491. __u8 set;
  492. /* Save current bank */
  493. set = inb(iobase + SSR);
  494. switch_bank(iobase, SET0);
  495. if (!(inb_p(iobase + USR) & USR_TSRE)) {
  496. pr_debug("%s: warning, FIFO not empty yet!\n", __func__);
  497. fifo_size -= 17;
  498. pr_debug("%s: %d bytes left in tx fifo\n", __func__, fifo_size);
  499. }
  500. /* Fill FIFO with current frame */
  501. while ((fifo_size-- > 0) && (actual < len)) {
  502. /* Transmit next byte */
  503. outb(buf[actual++], iobase + TBR);
  504. }
  505. pr_debug("%s: fifo_size %d ; %d sent of %d\n",
  506. __func__, fifo_size, actual, len);
  507. /* Restore bank */
  508. outb(set, iobase + SSR);
  509. return actual;
  510. }
  511. /*
  512. * Function w83977af_dma_xmit_complete (self)
  513. *
  514. * The transfer of a frame in finished. So do the necessary things
  515. *
  516. *
  517. */
  518. static void w83977af_dma_xmit_complete(struct w83977af_ir *self)
  519. {
  520. int iobase;
  521. __u8 set;
  522. pr_debug("%s: %ld\n", __func__, jiffies);
  523. IRDA_ASSERT(self, return;);
  524. iobase = self->io.fir_base;
  525. /* Save current set */
  526. set = inb(iobase + SSR);
  527. /* Disable DMA */
  528. switch_bank(iobase, SET0);
  529. outb(inb(iobase + HCR) & ~HCR_EN_DMA, iobase + HCR);
  530. /* Check for underrun! */
  531. if (inb(iobase + AUDR) & AUDR_UNDR) {
  532. pr_debug("%s: Transmit underrun!\n", __func__);
  533. self->netdev->stats.tx_errors++;
  534. self->netdev->stats.tx_fifo_errors++;
  535. /* Clear bit, by writing 1 to it */
  536. outb(AUDR_UNDR, iobase + AUDR);
  537. } else {
  538. self->netdev->stats.tx_packets++;
  539. }
  540. if (self->new_speed) {
  541. w83977af_change_speed(self, self->new_speed);
  542. self->new_speed = 0;
  543. }
  544. /* Unlock tx_buff and request another frame */
  545. /* Tell the network layer, that we want more frames */
  546. netif_wake_queue(self->netdev);
  547. /* Restore set */
  548. outb(set, iobase + SSR);
  549. }
  550. /*
  551. * Function w83977af_dma_receive (self)
  552. *
  553. * Get ready for receiving a frame. The device will initiate a DMA
  554. * if it starts to receive a frame.
  555. *
  556. */
  557. static int w83977af_dma_receive(struct w83977af_ir *self)
  558. {
  559. int iobase;
  560. __u8 set;
  561. #ifdef CONFIG_ARCH_NETWINDER
  562. unsigned long flags;
  563. __u8 hcr;
  564. #endif
  565. IRDA_ASSERT(self, return -1;);
  566. pr_debug("%s\n", __func__);
  567. iobase = self->io.fir_base;
  568. /* Save current set */
  569. set = inb(iobase + SSR);
  570. /* Disable DMA */
  571. switch_bank(iobase, SET0);
  572. outb(inb(iobase + HCR) & ~HCR_EN_DMA, iobase + HCR);
  573. /* Choose DMA Rx, DMA Fairness, and Advanced mode */
  574. switch_bank(iobase, SET2);
  575. outb((inb(iobase + ADCR1) & ~ADCR1_D_CHSW)/*|ADCR1_DMA_F*/ | ADCR1_ADV_SL,
  576. iobase + ADCR1);
  577. self->io.direction = IO_RECV;
  578. self->rx_buff.data = self->rx_buff.head;
  579. #ifdef CONFIG_ARCH_NETWINDER
  580. spin_lock_irqsave(&self->lock, flags);
  581. disable_dma(self->io.dma);
  582. clear_dma_ff(self->io.dma);
  583. set_dma_mode(self->io.dma, DMA_MODE_READ);
  584. set_dma_addr(self->io.dma, self->rx_buff_dma);
  585. set_dma_count(self->io.dma, self->rx_buff.truesize);
  586. #else
  587. irda_setup_dma(self->io.dma, self->rx_buff_dma, self->rx_buff.truesize,
  588. DMA_MODE_READ);
  589. #endif
  590. /*
  591. * Reset Rx FIFO. This will also flush the ST_FIFO, it's very
  592. * important that we don't reset the Tx FIFO since it might not
  593. * be finished transmitting yet
  594. */
  595. switch_bank(iobase, SET0);
  596. outb(UFR_RXTL | UFR_TXTL | UFR_RXF_RST | UFR_EN_FIFO, iobase + UFR);
  597. self->st_fifo.len = self->st_fifo.tail = self->st_fifo.head = 0;
  598. /* Enable DMA */
  599. switch_bank(iobase, SET0);
  600. #ifdef CONFIG_ARCH_NETWINDER
  601. hcr = inb(iobase + HCR);
  602. outb(hcr | HCR_EN_DMA, iobase + HCR);
  603. enable_dma(self->io.dma);
  604. spin_unlock_irqrestore(&self->lock, flags);
  605. #else
  606. outb(inb(iobase + HCR) | HCR_EN_DMA, iobase + HCR);
  607. #endif
  608. /* Restore set */
  609. outb(set, iobase + SSR);
  610. return 0;
  611. }
  612. /*
  613. * Function w83977af_receive_complete (self)
  614. *
  615. * Finished with receiving a frame
  616. *
  617. */
  618. static int w83977af_dma_receive_complete(struct w83977af_ir *self)
  619. {
  620. struct sk_buff *skb;
  621. struct st_fifo *st_fifo;
  622. int len;
  623. int iobase;
  624. __u8 set;
  625. __u8 status;
  626. pr_debug("%s\n", __func__);
  627. st_fifo = &self->st_fifo;
  628. iobase = self->io.fir_base;
  629. /* Save current set */
  630. set = inb(iobase + SSR);
  631. iobase = self->io.fir_base;
  632. /* Read status FIFO */
  633. switch_bank(iobase, SET5);
  634. while ((status = inb(iobase + FS_FO)) & FS_FO_FSFDR) {
  635. st_fifo->entries[st_fifo->tail].status = status;
  636. st_fifo->entries[st_fifo->tail].len = inb(iobase + RFLFL);
  637. st_fifo->entries[st_fifo->tail].len |= inb(iobase + RFLFH) << 8;
  638. st_fifo->tail++;
  639. st_fifo->len++;
  640. }
  641. while (st_fifo->len) {
  642. /* Get first entry */
  643. status = st_fifo->entries[st_fifo->head].status;
  644. len = st_fifo->entries[st_fifo->head].len;
  645. st_fifo->head++;
  646. st_fifo->len--;
  647. /* Check for errors */
  648. if (status & FS_FO_ERR_MSK) {
  649. if (status & FS_FO_LST_FR) {
  650. /* Add number of lost frames to stats */
  651. self->netdev->stats.rx_errors += len;
  652. } else {
  653. /* Skip frame */
  654. self->netdev->stats.rx_errors++;
  655. self->rx_buff.data += len;
  656. if (status & FS_FO_MX_LEX)
  657. self->netdev->stats.rx_length_errors++;
  658. if (status & FS_FO_PHY_ERR)
  659. self->netdev->stats.rx_frame_errors++;
  660. if (status & FS_FO_CRC_ERR)
  661. self->netdev->stats.rx_crc_errors++;
  662. }
  663. /* The errors below can be reported in both cases */
  664. if (status & FS_FO_RX_OV)
  665. self->netdev->stats.rx_fifo_errors++;
  666. if (status & FS_FO_FSF_OV)
  667. self->netdev->stats.rx_fifo_errors++;
  668. } else {
  669. /* Check if we have transferred all data to memory */
  670. switch_bank(iobase, SET0);
  671. if (inb(iobase + USR) & USR_RDR)
  672. udelay(80); /* Should be enough!? */
  673. skb = dev_alloc_skb(len + 1);
  674. if (!skb) {
  675. pr_info("%s: memory squeeze, dropping frame\n",
  676. __func__);
  677. /* Restore set register */
  678. outb(set, iobase + SSR);
  679. return FALSE;
  680. }
  681. /* Align to 20 bytes */
  682. skb_reserve(skb, 1);
  683. /* Copy frame without CRC */
  684. if (self->io.speed < 4000000) {
  685. skb_put(skb, len - 2);
  686. skb_copy_to_linear_data(skb,
  687. self->rx_buff.data,
  688. len - 2);
  689. } else {
  690. skb_put(skb, len - 4);
  691. skb_copy_to_linear_data(skb,
  692. self->rx_buff.data,
  693. len - 4);
  694. }
  695. /* Move to next frame */
  696. self->rx_buff.data += len;
  697. self->netdev->stats.rx_packets++;
  698. skb->dev = self->netdev;
  699. skb_reset_mac_header(skb);
  700. skb->protocol = htons(ETH_P_IRDA);
  701. netif_rx(skb);
  702. }
  703. }
  704. /* Restore set register */
  705. outb(set, iobase + SSR);
  706. return TRUE;
  707. }
  708. /*
  709. * Function pc87108_pio_receive (self)
  710. *
  711. * Receive all data in receiver FIFO
  712. *
  713. */
  714. static void w83977af_pio_receive(struct w83977af_ir *self)
  715. {
  716. __u8 byte = 0x00;
  717. int iobase;
  718. IRDA_ASSERT(self, return;);
  719. iobase = self->io.fir_base;
  720. /* Receive all characters in Rx FIFO */
  721. do {
  722. byte = inb(iobase + RBR);
  723. async_unwrap_char(self->netdev, &self->netdev->stats, &self->rx_buff,
  724. byte);
  725. } while (inb(iobase + USR) & USR_RDR); /* Data available */
  726. }
  727. /*
  728. * Function w83977af_sir_interrupt (self, eir)
  729. *
  730. * Handle SIR interrupt
  731. *
  732. */
  733. static __u8 w83977af_sir_interrupt(struct w83977af_ir *self, int isr)
  734. {
  735. int actual;
  736. __u8 new_icr = 0;
  737. __u8 set;
  738. int iobase;
  739. pr_debug("%s: isr=%#x\n", __func__, isr);
  740. iobase = self->io.fir_base;
  741. /* Transmit FIFO low on data */
  742. if (isr & ISR_TXTH_I) {
  743. /* Write data left in transmit buffer */
  744. actual = w83977af_pio_write(self->io.fir_base,
  745. self->tx_buff.data,
  746. self->tx_buff.len,
  747. self->io.fifo_size);
  748. self->tx_buff.data += actual;
  749. self->tx_buff.len -= actual;
  750. self->io.direction = IO_XMIT;
  751. /* Check if finished */
  752. if (self->tx_buff.len > 0) {
  753. new_icr |= ICR_ETXTHI;
  754. } else {
  755. set = inb(iobase + SSR);
  756. switch_bank(iobase, SET0);
  757. outb(AUDR_SFEND, iobase + AUDR);
  758. outb(set, iobase + SSR);
  759. self->netdev->stats.tx_packets++;
  760. /* Feed me more packets */
  761. netif_wake_queue(self->netdev);
  762. new_icr |= ICR_ETBREI;
  763. }
  764. }
  765. /* Check if transmission has completed */
  766. if (isr & ISR_TXEMP_I) {
  767. /* Check if we need to change the speed? */
  768. if (self->new_speed) {
  769. pr_debug("%s: Changing speed!\n", __func__);
  770. w83977af_change_speed(self, self->new_speed);
  771. self->new_speed = 0;
  772. }
  773. /* Turn around and get ready to receive some data */
  774. self->io.direction = IO_RECV;
  775. new_icr |= ICR_ERBRI;
  776. }
  777. /* Rx FIFO threshold or timeout */
  778. if (isr & ISR_RXTH_I) {
  779. w83977af_pio_receive(self);
  780. /* Keep receiving */
  781. new_icr |= ICR_ERBRI;
  782. }
  783. return new_icr;
  784. }
  785. /*
  786. * Function pc87108_fir_interrupt (self, eir)
  787. *
  788. * Handle MIR/FIR interrupt
  789. *
  790. */
  791. static __u8 w83977af_fir_interrupt(struct w83977af_ir *self, int isr)
  792. {
  793. __u8 new_icr = 0;
  794. __u8 set;
  795. int iobase;
  796. iobase = self->io.fir_base;
  797. set = inb(iobase + SSR);
  798. /* End of frame detected in FIFO */
  799. if (isr & (ISR_FEND_I | ISR_FSF_I)) {
  800. if (w83977af_dma_receive_complete(self)) {
  801. /* Wait for next status FIFO interrupt */
  802. new_icr |= ICR_EFSFI;
  803. } else {
  804. /* DMA not finished yet */
  805. /* Set timer value, resolution 1 ms */
  806. switch_bank(iobase, SET4);
  807. outb(0x01, iobase + TMRL); /* 1 ms */
  808. outb(0x00, iobase + TMRH);
  809. /* Start timer */
  810. outb(IR_MSL_EN_TMR, iobase + IR_MSL);
  811. new_icr |= ICR_ETMRI;
  812. }
  813. }
  814. /* Timer finished */
  815. if (isr & ISR_TMR_I) {
  816. /* Disable timer */
  817. switch_bank(iobase, SET4);
  818. outb(0, iobase + IR_MSL);
  819. /* Clear timer event */
  820. /* switch_bank(iobase, SET0); */
  821. /* outb(ASCR_CTE, iobase+ASCR); */
  822. /* Check if this is a TX timer interrupt */
  823. if (self->io.direction == IO_XMIT) {
  824. w83977af_dma_write(self, iobase);
  825. new_icr |= ICR_EDMAI;
  826. } else {
  827. /* Check if DMA has now finished */
  828. w83977af_dma_receive_complete(self);
  829. new_icr |= ICR_EFSFI;
  830. }
  831. }
  832. /* Finished with DMA */
  833. if (isr & ISR_DMA_I) {
  834. w83977af_dma_xmit_complete(self);
  835. /* Check if there are more frames to be transmitted */
  836. /* if (irda_device_txqueue_empty(self)) { */
  837. /* Prepare for receive
  838. *
  839. * ** Netwinder Tx DMA likes that we do this anyway **
  840. */
  841. w83977af_dma_receive(self);
  842. new_icr = ICR_EFSFI;
  843. /* } */
  844. }
  845. /* Restore set */
  846. outb(set, iobase + SSR);
  847. return new_icr;
  848. }
  849. /*
  850. * Function w83977af_interrupt (irq, dev_id, regs)
  851. *
  852. * An interrupt from the chip has arrived. Time to do some work
  853. *
  854. */
  855. static irqreturn_t w83977af_interrupt(int irq, void *dev_id)
  856. {
  857. struct net_device *dev = dev_id;
  858. struct w83977af_ir *self;
  859. __u8 set, icr, isr;
  860. int iobase;
  861. self = netdev_priv(dev);
  862. iobase = self->io.fir_base;
  863. /* Save current bank */
  864. set = inb(iobase + SSR);
  865. switch_bank(iobase, SET0);
  866. icr = inb(iobase + ICR);
  867. isr = inb(iobase + ISR) & icr; /* Mask out the interesting ones */
  868. outb(0, iobase + ICR); /* Disable interrupts */
  869. if (isr) {
  870. /* Dispatch interrupt handler for the current speed */
  871. if (self->io.speed > PIO_MAX_SPEED)
  872. icr = w83977af_fir_interrupt(self, isr);
  873. else
  874. icr = w83977af_sir_interrupt(self, isr);
  875. }
  876. outb(icr, iobase + ICR); /* Restore (new) interrupts */
  877. outb(set, iobase + SSR); /* Restore bank register */
  878. return IRQ_RETVAL(isr);
  879. }
  880. /*
  881. * Function w83977af_is_receiving (self)
  882. *
  883. * Return TRUE is we are currently receiving a frame
  884. *
  885. */
  886. static int w83977af_is_receiving(struct w83977af_ir *self)
  887. {
  888. int status = FALSE;
  889. int iobase;
  890. __u8 set;
  891. IRDA_ASSERT(self, return FALSE;);
  892. if (self->io.speed > 115200) {
  893. iobase = self->io.fir_base;
  894. /* Check if rx FIFO is not empty */
  895. set = inb(iobase + SSR);
  896. switch_bank(iobase, SET2);
  897. if ((inb(iobase + RXFDTH) & 0x3f) != 0) {
  898. /* We are receiving something */
  899. status = TRUE;
  900. }
  901. outb(set, iobase + SSR);
  902. } else {
  903. status = (self->rx_buff.state != OUTSIDE_FRAME);
  904. }
  905. return status;
  906. }
  907. /*
  908. * Function w83977af_net_open (dev)
  909. *
  910. * Start the device
  911. *
  912. */
  913. static int w83977af_net_open(struct net_device *dev)
  914. {
  915. struct w83977af_ir *self;
  916. int iobase;
  917. char hwname[32];
  918. __u8 set;
  919. IRDA_ASSERT(dev, return -1;);
  920. self = netdev_priv(dev);
  921. IRDA_ASSERT(self, return 0;);
  922. iobase = self->io.fir_base;
  923. if (request_irq(self->io.irq, w83977af_interrupt, 0, dev->name,
  924. (void *)dev)) {
  925. return -EAGAIN;
  926. }
  927. /*
  928. * Always allocate the DMA channel after the IRQ,
  929. * and clean up on failure.
  930. */
  931. if (request_dma(self->io.dma, dev->name)) {
  932. free_irq(self->io.irq, dev);
  933. return -EAGAIN;
  934. }
  935. /* Save current set */
  936. set = inb(iobase + SSR);
  937. /* Enable some interrupts so we can receive frames again */
  938. switch_bank(iobase, SET0);
  939. if (self->io.speed > 115200) {
  940. outb(ICR_EFSFI, iobase + ICR);
  941. w83977af_dma_receive(self);
  942. } else {
  943. outb(ICR_ERBRI, iobase + ICR);
  944. }
  945. /* Restore bank register */
  946. outb(set, iobase + SSR);
  947. /* Ready to play! */
  948. netif_start_queue(dev);
  949. /* Give self a hardware name */
  950. sprintf(hwname, "w83977af @ 0x%03x", self->io.fir_base);
  951. /*
  952. * Open new IrLAP layer instance, now that everything should be
  953. * initialized properly
  954. */
  955. self->irlap = irlap_open(dev, &self->qos, hwname);
  956. return 0;
  957. }
  958. /*
  959. * Function w83977af_net_close (dev)
  960. *
  961. * Stop the device
  962. *
  963. */
  964. static int w83977af_net_close(struct net_device *dev)
  965. {
  966. struct w83977af_ir *self;
  967. int iobase;
  968. __u8 set;
  969. IRDA_ASSERT(dev, return -1;);
  970. self = netdev_priv(dev);
  971. IRDA_ASSERT(self, return 0;);
  972. iobase = self->io.fir_base;
  973. /* Stop device */
  974. netif_stop_queue(dev);
  975. /* Stop and remove instance of IrLAP */
  976. if (self->irlap)
  977. irlap_close(self->irlap);
  978. self->irlap = NULL;
  979. disable_dma(self->io.dma);
  980. /* Save current set */
  981. set = inb(iobase + SSR);
  982. /* Disable interrupts */
  983. switch_bank(iobase, SET0);
  984. outb(0, iobase + ICR);
  985. free_irq(self->io.irq, dev);
  986. free_dma(self->io.dma);
  987. /* Restore bank register */
  988. outb(set, iobase + SSR);
  989. return 0;
  990. }
  991. /*
  992. * Function w83977af_net_ioctl (dev, rq, cmd)
  993. *
  994. * Process IOCTL commands for this device
  995. *
  996. */
  997. static int w83977af_net_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
  998. {
  999. struct if_irda_req *irq = (struct if_irda_req *)rq;
  1000. struct w83977af_ir *self;
  1001. unsigned long flags;
  1002. int ret = 0;
  1003. IRDA_ASSERT(dev, return -1;);
  1004. self = netdev_priv(dev);
  1005. IRDA_ASSERT(self, return -1;);
  1006. pr_debug("%s: %s, (cmd=0x%X)\n", __func__, dev->name, cmd);
  1007. spin_lock_irqsave(&self->lock, flags);
  1008. switch (cmd) {
  1009. case SIOCSBANDWIDTH: /* Set bandwidth */
  1010. if (!capable(CAP_NET_ADMIN)) {
  1011. ret = -EPERM;
  1012. goto out;
  1013. }
  1014. w83977af_change_speed(self, irq->ifr_baudrate);
  1015. break;
  1016. case SIOCSMEDIABUSY: /* Set media busy */
  1017. if (!capable(CAP_NET_ADMIN)) {
  1018. ret = -EPERM;
  1019. goto out;
  1020. }
  1021. irda_device_set_media_busy(self->netdev, TRUE);
  1022. break;
  1023. case SIOCGRECEIVING: /* Check if we are receiving right now */
  1024. irq->ifr_receiving = w83977af_is_receiving(self);
  1025. break;
  1026. default:
  1027. ret = -EOPNOTSUPP;
  1028. }
  1029. out:
  1030. spin_unlock_irqrestore(&self->lock, flags);
  1031. return ret;
  1032. }
  1033. MODULE_AUTHOR("Dag Brattli <dagb@cs.uit.no>");
  1034. MODULE_DESCRIPTION("Winbond W83977AF IrDA Device Driver");
  1035. MODULE_LICENSE("GPL");
  1036. module_param(qos_mtt_bits, int, 0);
  1037. MODULE_PARM_DESC(qos_mtt_bits, "Mimimum Turn Time");
  1038. module_param_array(io, int, NULL, 0);
  1039. MODULE_PARM_DESC(io, "Base I/O addresses");
  1040. module_param_array(irq, int, NULL, 0);
  1041. MODULE_PARM_DESC(irq, "IRQ lines");
  1042. /*
  1043. * Function init_module (void)
  1044. *
  1045. *
  1046. *
  1047. */
  1048. module_init(w83977af_init);
  1049. /*
  1050. * Function cleanup_module (void)
  1051. *
  1052. *
  1053. *
  1054. */
  1055. module_exit(w83977af_cleanup);