mpsc.c 55 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723172417251726172717281729173017311732173317341735173617371738173917401741174217431744174517461747174817491750175117521753175417551756175717581759176017611762176317641765176617671768176917701771177217731774177517761777177817791780178117821783178417851786178717881789179017911792179317941795179617971798179918001801180218031804180518061807180818091810181118121813181418151816181718181819182018211822182318241825182618271828182918301831183218331834183518361837183818391840184118421843184418451846184718481849185018511852185318541855185618571858185918601861186218631864186518661867186818691870187118721873187418751876187718781879188018811882188318841885188618871888188918901891189218931894189518961897189818991900190119021903190419051906190719081909191019111912191319141915191619171918191919201921192219231924192519261927192819291930193119321933193419351936193719381939194019411942194319441945194619471948194919501951195219531954195519561957195819591960196119621963196419651966196719681969197019711972197319741975197619771978197919801981198219831984198519861987198819891990199119921993199419951996199719981999200020012002200320042005200620072008200920102011201220132014201520162017201820192020202120222023202420252026202720282029203020312032203320342035203620372038203920402041204220432044204520462047204820492050205120522053205420552056205720582059206020612062206320642065206620672068206920702071207220732074207520762077207820792080208120822083208420852086208720882089209020912092209320942095209620972098209921002101210221032104210521062107210821092110211121122113211421152116211721182119212021212122212321242125212621272128212921302131213221332134213521362137213821392140
  1. /*
  2. * Generic driver for the MPSC (UART mode) on Marvell parts (e.g., GT64240,
  3. * GT64260, MV64340, MV64360, GT96100, ... ).
  4. *
  5. * Author: Mark A. Greer <mgreer@mvista.com>
  6. *
  7. * Based on an old MPSC driver that was in the linuxppc tree. It appears to
  8. * have been created by Chris Zankel (formerly of MontaVista) but there
  9. * is no proper Copyright so I'm not sure. Apparently, parts were also
  10. * taken from PPCBoot (now U-Boot). Also based on drivers/serial/8250.c
  11. * by Russell King.
  12. *
  13. * 2004 (c) MontaVista, Software, Inc. This file is licensed under
  14. * the terms of the GNU General Public License version 2. This program
  15. * is licensed "as is" without any warranty of any kind, whether express
  16. * or implied.
  17. */
  18. /*
  19. * The MPSC interface is much like a typical network controller's interface.
  20. * That is, you set up separate rings of descriptors for transmitting and
  21. * receiving data. There is also a pool of buffers with (one buffer per
  22. * descriptor) that incoming data are dma'd into or outgoing data are dma'd
  23. * out of.
  24. *
  25. * The MPSC requires two other controllers to be able to work. The Baud Rate
  26. * Generator (BRG) provides a clock at programmable frequencies which determines
  27. * the baud rate. The Serial DMA Controller (SDMA) takes incoming data from the
  28. * MPSC and DMA's it into memory or DMA's outgoing data and passes it to the
  29. * MPSC. It is actually the SDMA interrupt that the driver uses to keep the
  30. * transmit and receive "engines" going (i.e., indicate data has been
  31. * transmitted or received).
  32. *
  33. * NOTES:
  34. *
  35. * 1) Some chips have an erratum where several regs cannot be
  36. * read. To work around that, we keep a local copy of those regs in
  37. * 'mpsc_port_info'.
  38. *
  39. * 2) Some chips have an erratum where the ctlr will hang when the SDMA ctlr
  40. * accesses system mem with coherency enabled. For that reason, the driver
  41. * assumes that coherency for that ctlr has been disabled. This means
  42. * that when in a cache coherent system, the driver has to manually manage
  43. * the data cache on the areas that it touches because the dma_* macro are
  44. * basically no-ops.
  45. *
  46. * 3) There is an erratum (on PPC) where you can't use the instruction to do
  47. * a DMA_TO_DEVICE/cache clean so DMA_BIDIRECTIONAL/flushes are used in places
  48. * where a DMA_TO_DEVICE/clean would have [otherwise] sufficed.
  49. *
  50. * 4) AFAICT, hardware flow control isn't supported by the controller --MAG.
  51. */
  52. #if defined(CONFIG_SERIAL_MPSC_CONSOLE) && defined(CONFIG_MAGIC_SYSRQ)
  53. #define SUPPORT_SYSRQ
  54. #endif
  55. #include <linux/tty.h>
  56. #include <linux/tty_flip.h>
  57. #include <linux/ioport.h>
  58. #include <linux/init.h>
  59. #include <linux/console.h>
  60. #include <linux/sysrq.h>
  61. #include <linux/serial.h>
  62. #include <linux/serial_core.h>
  63. #include <linux/delay.h>
  64. #include <linux/device.h>
  65. #include <linux/dma-mapping.h>
  66. #include <linux/mv643xx.h>
  67. #include <linux/platform_device.h>
  68. #include <linux/gfp.h>
  69. #include <asm/io.h>
  70. #include <asm/irq.h>
  71. #define MPSC_NUM_CTLRS 2
  72. /*
  73. * Descriptors and buffers must be cache line aligned.
  74. * Buffers lengths must be multiple of cache line size.
  75. * Number of Tx & Rx descriptors must be powers of 2.
  76. */
  77. #define MPSC_RXR_ENTRIES 32
  78. #define MPSC_RXRE_SIZE dma_get_cache_alignment()
  79. #define MPSC_RXR_SIZE (MPSC_RXR_ENTRIES * MPSC_RXRE_SIZE)
  80. #define MPSC_RXBE_SIZE dma_get_cache_alignment()
  81. #define MPSC_RXB_SIZE (MPSC_RXR_ENTRIES * MPSC_RXBE_SIZE)
  82. #define MPSC_TXR_ENTRIES 32
  83. #define MPSC_TXRE_SIZE dma_get_cache_alignment()
  84. #define MPSC_TXR_SIZE (MPSC_TXR_ENTRIES * MPSC_TXRE_SIZE)
  85. #define MPSC_TXBE_SIZE dma_get_cache_alignment()
  86. #define MPSC_TXB_SIZE (MPSC_TXR_ENTRIES * MPSC_TXBE_SIZE)
  87. #define MPSC_DMA_ALLOC_SIZE (MPSC_RXR_SIZE + MPSC_RXB_SIZE + MPSC_TXR_SIZE \
  88. + MPSC_TXB_SIZE + dma_get_cache_alignment() /* for alignment */)
  89. /* Rx and Tx Ring entry descriptors -- assume entry size is <= cacheline size */
  90. struct mpsc_rx_desc {
  91. u16 bufsize;
  92. u16 bytecnt;
  93. u32 cmdstat;
  94. u32 link;
  95. u32 buf_ptr;
  96. } __attribute((packed));
  97. struct mpsc_tx_desc {
  98. u16 bytecnt;
  99. u16 shadow;
  100. u32 cmdstat;
  101. u32 link;
  102. u32 buf_ptr;
  103. } __attribute((packed));
  104. /*
  105. * Some regs that have the erratum that you can't read them are are shared
  106. * between the two MPSC controllers. This struct contains those shared regs.
  107. */
  108. struct mpsc_shared_regs {
  109. phys_addr_t mpsc_routing_base_p;
  110. phys_addr_t sdma_intr_base_p;
  111. void __iomem *mpsc_routing_base;
  112. void __iomem *sdma_intr_base;
  113. u32 MPSC_MRR_m;
  114. u32 MPSC_RCRR_m;
  115. u32 MPSC_TCRR_m;
  116. u32 SDMA_INTR_CAUSE_m;
  117. u32 SDMA_INTR_MASK_m;
  118. };
  119. /* The main driver data structure */
  120. struct mpsc_port_info {
  121. struct uart_port port; /* Overlay uart_port structure */
  122. /* Internal driver state for this ctlr */
  123. u8 ready;
  124. u8 rcv_data;
  125. /* Info passed in from platform */
  126. u8 mirror_regs; /* Need to mirror regs? */
  127. u8 cache_mgmt; /* Need manual cache mgmt? */
  128. u8 brg_can_tune; /* BRG has baud tuning? */
  129. u32 brg_clk_src;
  130. u16 mpsc_max_idle;
  131. int default_baud;
  132. int default_bits;
  133. int default_parity;
  134. int default_flow;
  135. /* Physical addresses of various blocks of registers (from platform) */
  136. phys_addr_t mpsc_base_p;
  137. phys_addr_t sdma_base_p;
  138. phys_addr_t brg_base_p;
  139. /* Virtual addresses of various blocks of registers (from platform) */
  140. void __iomem *mpsc_base;
  141. void __iomem *sdma_base;
  142. void __iomem *brg_base;
  143. /* Descriptor ring and buffer allocations */
  144. void *dma_region;
  145. dma_addr_t dma_region_p;
  146. dma_addr_t rxr; /* Rx descriptor ring */
  147. dma_addr_t rxr_p; /* Phys addr of rxr */
  148. u8 *rxb; /* Rx Ring I/O buf */
  149. u8 *rxb_p; /* Phys addr of rxb */
  150. u32 rxr_posn; /* First desc w/ Rx data */
  151. dma_addr_t txr; /* Tx descriptor ring */
  152. dma_addr_t txr_p; /* Phys addr of txr */
  153. u8 *txb; /* Tx Ring I/O buf */
  154. u8 *txb_p; /* Phys addr of txb */
  155. int txr_head; /* Where new data goes */
  156. int txr_tail; /* Where sent data comes off */
  157. spinlock_t tx_lock; /* transmit lock */
  158. /* Mirrored values of regs we can't read (if 'mirror_regs' set) */
  159. u32 MPSC_MPCR_m;
  160. u32 MPSC_CHR_1_m;
  161. u32 MPSC_CHR_2_m;
  162. u32 MPSC_CHR_10_m;
  163. u32 BRG_BCR_m;
  164. struct mpsc_shared_regs *shared_regs;
  165. };
  166. /* Hooks to platform-specific code */
  167. int mpsc_platform_register_driver(void);
  168. void mpsc_platform_unregister_driver(void);
  169. /* Hooks back in to mpsc common to be called by platform-specific code */
  170. struct mpsc_port_info *mpsc_device_probe(int index);
  171. struct mpsc_port_info *mpsc_device_remove(int index);
  172. /* Main MPSC Configuration Register Offsets */
  173. #define MPSC_MMCRL 0x0000
  174. #define MPSC_MMCRH 0x0004
  175. #define MPSC_MPCR 0x0008
  176. #define MPSC_CHR_1 0x000c
  177. #define MPSC_CHR_2 0x0010
  178. #define MPSC_CHR_3 0x0014
  179. #define MPSC_CHR_4 0x0018
  180. #define MPSC_CHR_5 0x001c
  181. #define MPSC_CHR_6 0x0020
  182. #define MPSC_CHR_7 0x0024
  183. #define MPSC_CHR_8 0x0028
  184. #define MPSC_CHR_9 0x002c
  185. #define MPSC_CHR_10 0x0030
  186. #define MPSC_CHR_11 0x0034
  187. #define MPSC_MPCR_FRZ (1 << 9)
  188. #define MPSC_MPCR_CL_5 0
  189. #define MPSC_MPCR_CL_6 1
  190. #define MPSC_MPCR_CL_7 2
  191. #define MPSC_MPCR_CL_8 3
  192. #define MPSC_MPCR_SBL_1 0
  193. #define MPSC_MPCR_SBL_2 1
  194. #define MPSC_CHR_2_TEV (1<<1)
  195. #define MPSC_CHR_2_TA (1<<7)
  196. #define MPSC_CHR_2_TTCS (1<<9)
  197. #define MPSC_CHR_2_REV (1<<17)
  198. #define MPSC_CHR_2_RA (1<<23)
  199. #define MPSC_CHR_2_CRD (1<<25)
  200. #define MPSC_CHR_2_EH (1<<31)
  201. #define MPSC_CHR_2_PAR_ODD 0
  202. #define MPSC_CHR_2_PAR_SPACE 1
  203. #define MPSC_CHR_2_PAR_EVEN 2
  204. #define MPSC_CHR_2_PAR_MARK 3
  205. /* MPSC Signal Routing */
  206. #define MPSC_MRR 0x0000
  207. #define MPSC_RCRR 0x0004
  208. #define MPSC_TCRR 0x0008
  209. /* Serial DMA Controller Interface Registers */
  210. #define SDMA_SDC 0x0000
  211. #define SDMA_SDCM 0x0008
  212. #define SDMA_RX_DESC 0x0800
  213. #define SDMA_RX_BUF_PTR 0x0808
  214. #define SDMA_SCRDP 0x0810
  215. #define SDMA_TX_DESC 0x0c00
  216. #define SDMA_SCTDP 0x0c10
  217. #define SDMA_SFTDP 0x0c14
  218. #define SDMA_DESC_CMDSTAT_PE (1<<0)
  219. #define SDMA_DESC_CMDSTAT_CDL (1<<1)
  220. #define SDMA_DESC_CMDSTAT_FR (1<<3)
  221. #define SDMA_DESC_CMDSTAT_OR (1<<6)
  222. #define SDMA_DESC_CMDSTAT_BR (1<<9)
  223. #define SDMA_DESC_CMDSTAT_MI (1<<10)
  224. #define SDMA_DESC_CMDSTAT_A (1<<11)
  225. #define SDMA_DESC_CMDSTAT_AM (1<<12)
  226. #define SDMA_DESC_CMDSTAT_CT (1<<13)
  227. #define SDMA_DESC_CMDSTAT_C (1<<14)
  228. #define SDMA_DESC_CMDSTAT_ES (1<<15)
  229. #define SDMA_DESC_CMDSTAT_L (1<<16)
  230. #define SDMA_DESC_CMDSTAT_F (1<<17)
  231. #define SDMA_DESC_CMDSTAT_P (1<<18)
  232. #define SDMA_DESC_CMDSTAT_EI (1<<23)
  233. #define SDMA_DESC_CMDSTAT_O (1<<31)
  234. #define SDMA_DESC_DFLT (SDMA_DESC_CMDSTAT_O \
  235. | SDMA_DESC_CMDSTAT_EI)
  236. #define SDMA_SDC_RFT (1<<0)
  237. #define SDMA_SDC_SFM (1<<1)
  238. #define SDMA_SDC_BLMR (1<<6)
  239. #define SDMA_SDC_BLMT (1<<7)
  240. #define SDMA_SDC_POVR (1<<8)
  241. #define SDMA_SDC_RIFB (1<<9)
  242. #define SDMA_SDCM_ERD (1<<7)
  243. #define SDMA_SDCM_AR (1<<15)
  244. #define SDMA_SDCM_STD (1<<16)
  245. #define SDMA_SDCM_TXD (1<<23)
  246. #define SDMA_SDCM_AT (1<<31)
  247. #define SDMA_0_CAUSE_RXBUF (1<<0)
  248. #define SDMA_0_CAUSE_RXERR (1<<1)
  249. #define SDMA_0_CAUSE_TXBUF (1<<2)
  250. #define SDMA_0_CAUSE_TXEND (1<<3)
  251. #define SDMA_1_CAUSE_RXBUF (1<<8)
  252. #define SDMA_1_CAUSE_RXERR (1<<9)
  253. #define SDMA_1_CAUSE_TXBUF (1<<10)
  254. #define SDMA_1_CAUSE_TXEND (1<<11)
  255. #define SDMA_CAUSE_RX_MASK (SDMA_0_CAUSE_RXBUF | SDMA_0_CAUSE_RXERR \
  256. | SDMA_1_CAUSE_RXBUF | SDMA_1_CAUSE_RXERR)
  257. #define SDMA_CAUSE_TX_MASK (SDMA_0_CAUSE_TXBUF | SDMA_0_CAUSE_TXEND \
  258. | SDMA_1_CAUSE_TXBUF | SDMA_1_CAUSE_TXEND)
  259. /* SDMA Interrupt registers */
  260. #define SDMA_INTR_CAUSE 0x0000
  261. #define SDMA_INTR_MASK 0x0080
  262. /* Baud Rate Generator Interface Registers */
  263. #define BRG_BCR 0x0000
  264. #define BRG_BTR 0x0004
  265. /*
  266. * Define how this driver is known to the outside (we've been assigned a
  267. * range on the "Low-density serial ports" major).
  268. */
  269. #define MPSC_MAJOR 204
  270. #define MPSC_MINOR_START 44
  271. #define MPSC_DRIVER_NAME "MPSC"
  272. #define MPSC_DEV_NAME "ttyMM"
  273. #define MPSC_VERSION "1.00"
  274. static struct mpsc_port_info mpsc_ports[MPSC_NUM_CTLRS];
  275. static struct mpsc_shared_regs mpsc_shared_regs;
  276. static struct uart_driver mpsc_reg;
  277. static void mpsc_start_rx(struct mpsc_port_info *pi);
  278. static void mpsc_free_ring_mem(struct mpsc_port_info *pi);
  279. static void mpsc_release_port(struct uart_port *port);
  280. /*
  281. ******************************************************************************
  282. *
  283. * Baud Rate Generator Routines (BRG)
  284. *
  285. ******************************************************************************
  286. */
  287. static void mpsc_brg_init(struct mpsc_port_info *pi, u32 clk_src)
  288. {
  289. u32 v;
  290. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  291. v = (v & ~(0xf << 18)) | ((clk_src & 0xf) << 18);
  292. if (pi->brg_can_tune)
  293. v &= ~(1 << 25);
  294. if (pi->mirror_regs)
  295. pi->BRG_BCR_m = v;
  296. writel(v, pi->brg_base + BRG_BCR);
  297. writel(readl(pi->brg_base + BRG_BTR) & 0xffff0000,
  298. pi->brg_base + BRG_BTR);
  299. }
  300. static void mpsc_brg_enable(struct mpsc_port_info *pi)
  301. {
  302. u32 v;
  303. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  304. v |= (1 << 16);
  305. if (pi->mirror_regs)
  306. pi->BRG_BCR_m = v;
  307. writel(v, pi->brg_base + BRG_BCR);
  308. }
  309. static void mpsc_brg_disable(struct mpsc_port_info *pi)
  310. {
  311. u32 v;
  312. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  313. v &= ~(1 << 16);
  314. if (pi->mirror_regs)
  315. pi->BRG_BCR_m = v;
  316. writel(v, pi->brg_base + BRG_BCR);
  317. }
  318. /*
  319. * To set the baud, we adjust the CDV field in the BRG_BCR reg.
  320. * From manual: Baud = clk / ((CDV+1)*2) ==> CDV = (clk / (baud*2)) - 1.
  321. * However, the input clock is divided by 16 in the MPSC b/c of how
  322. * 'MPSC_MMCRH' was set up so we have to divide the 'clk' used in our
  323. * calculation by 16 to account for that. So the real calculation
  324. * that accounts for the way the mpsc is set up is:
  325. * CDV = (clk / (baud*2*16)) - 1 ==> CDV = (clk / (baud << 5)) - 1.
  326. */
  327. static void mpsc_set_baudrate(struct mpsc_port_info *pi, u32 baud)
  328. {
  329. u32 cdv = (pi->port.uartclk / (baud << 5)) - 1;
  330. u32 v;
  331. mpsc_brg_disable(pi);
  332. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  333. v = (v & 0xffff0000) | (cdv & 0xffff);
  334. if (pi->mirror_regs)
  335. pi->BRG_BCR_m = v;
  336. writel(v, pi->brg_base + BRG_BCR);
  337. mpsc_brg_enable(pi);
  338. }
  339. /*
  340. ******************************************************************************
  341. *
  342. * Serial DMA Routines (SDMA)
  343. *
  344. ******************************************************************************
  345. */
  346. static void mpsc_sdma_burstsize(struct mpsc_port_info *pi, u32 burst_size)
  347. {
  348. u32 v;
  349. pr_debug("mpsc_sdma_burstsize[%d]: burst_size: %d\n",
  350. pi->port.line, burst_size);
  351. burst_size >>= 3; /* Divide by 8 b/c reg values are 8-byte chunks */
  352. if (burst_size < 2)
  353. v = 0x0; /* 1 64-bit word */
  354. else if (burst_size < 4)
  355. v = 0x1; /* 2 64-bit words */
  356. else if (burst_size < 8)
  357. v = 0x2; /* 4 64-bit words */
  358. else
  359. v = 0x3; /* 8 64-bit words */
  360. writel((readl(pi->sdma_base + SDMA_SDC) & (0x3 << 12)) | (v << 12),
  361. pi->sdma_base + SDMA_SDC);
  362. }
  363. static void mpsc_sdma_init(struct mpsc_port_info *pi, u32 burst_size)
  364. {
  365. pr_debug("mpsc_sdma_init[%d]: burst_size: %d\n", pi->port.line,
  366. burst_size);
  367. writel((readl(pi->sdma_base + SDMA_SDC) & 0x3ff) | 0x03f,
  368. pi->sdma_base + SDMA_SDC);
  369. mpsc_sdma_burstsize(pi, burst_size);
  370. }
  371. static u32 mpsc_sdma_intr_mask(struct mpsc_port_info *pi, u32 mask)
  372. {
  373. u32 old, v;
  374. pr_debug("mpsc_sdma_intr_mask[%d]: mask: 0x%x\n", pi->port.line, mask);
  375. old = v = (pi->mirror_regs) ? pi->shared_regs->SDMA_INTR_MASK_m :
  376. readl(pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  377. mask &= 0xf;
  378. if (pi->port.line)
  379. mask <<= 8;
  380. v &= ~mask;
  381. if (pi->mirror_regs)
  382. pi->shared_regs->SDMA_INTR_MASK_m = v;
  383. writel(v, pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  384. if (pi->port.line)
  385. old >>= 8;
  386. return old & 0xf;
  387. }
  388. static void mpsc_sdma_intr_unmask(struct mpsc_port_info *pi, u32 mask)
  389. {
  390. u32 v;
  391. pr_debug("mpsc_sdma_intr_unmask[%d]: mask: 0x%x\n", pi->port.line,mask);
  392. v = (pi->mirror_regs) ? pi->shared_regs->SDMA_INTR_MASK_m
  393. : readl(pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  394. mask &= 0xf;
  395. if (pi->port.line)
  396. mask <<= 8;
  397. v |= mask;
  398. if (pi->mirror_regs)
  399. pi->shared_regs->SDMA_INTR_MASK_m = v;
  400. writel(v, pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  401. }
  402. static void mpsc_sdma_intr_ack(struct mpsc_port_info *pi)
  403. {
  404. pr_debug("mpsc_sdma_intr_ack[%d]: Acknowledging IRQ\n", pi->port.line);
  405. if (pi->mirror_regs)
  406. pi->shared_regs->SDMA_INTR_CAUSE_m = 0;
  407. writeb(0x00, pi->shared_regs->sdma_intr_base + SDMA_INTR_CAUSE
  408. + pi->port.line);
  409. }
  410. static void mpsc_sdma_set_rx_ring(struct mpsc_port_info *pi,
  411. struct mpsc_rx_desc *rxre_p)
  412. {
  413. pr_debug("mpsc_sdma_set_rx_ring[%d]: rxre_p: 0x%x\n",
  414. pi->port.line, (u32)rxre_p);
  415. writel((u32)rxre_p, pi->sdma_base + SDMA_SCRDP);
  416. }
  417. static void mpsc_sdma_set_tx_ring(struct mpsc_port_info *pi,
  418. struct mpsc_tx_desc *txre_p)
  419. {
  420. writel((u32)txre_p, pi->sdma_base + SDMA_SFTDP);
  421. writel((u32)txre_p, pi->sdma_base + SDMA_SCTDP);
  422. }
  423. static void mpsc_sdma_cmd(struct mpsc_port_info *pi, u32 val)
  424. {
  425. u32 v;
  426. v = readl(pi->sdma_base + SDMA_SDCM);
  427. if (val)
  428. v |= val;
  429. else
  430. v = 0;
  431. wmb();
  432. writel(v, pi->sdma_base + SDMA_SDCM);
  433. wmb();
  434. }
  435. static uint mpsc_sdma_tx_active(struct mpsc_port_info *pi)
  436. {
  437. return readl(pi->sdma_base + SDMA_SDCM) & SDMA_SDCM_TXD;
  438. }
  439. static void mpsc_sdma_start_tx(struct mpsc_port_info *pi)
  440. {
  441. struct mpsc_tx_desc *txre, *txre_p;
  442. /* If tx isn't running & there's a desc ready to go, start it */
  443. if (!mpsc_sdma_tx_active(pi)) {
  444. txre = (struct mpsc_tx_desc *)(pi->txr
  445. + (pi->txr_tail * MPSC_TXRE_SIZE));
  446. dma_cache_sync(pi->port.dev, (void *)txre, MPSC_TXRE_SIZE,
  447. DMA_FROM_DEVICE);
  448. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  449. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  450. invalidate_dcache_range((ulong)txre,
  451. (ulong)txre + MPSC_TXRE_SIZE);
  452. #endif
  453. if (be32_to_cpu(txre->cmdstat) & SDMA_DESC_CMDSTAT_O) {
  454. txre_p = (struct mpsc_tx_desc *)
  455. (pi->txr_p + (pi->txr_tail * MPSC_TXRE_SIZE));
  456. mpsc_sdma_set_tx_ring(pi, txre_p);
  457. mpsc_sdma_cmd(pi, SDMA_SDCM_STD | SDMA_SDCM_TXD);
  458. }
  459. }
  460. }
  461. static void mpsc_sdma_stop(struct mpsc_port_info *pi)
  462. {
  463. pr_debug("mpsc_sdma_stop[%d]: Stopping SDMA\n", pi->port.line);
  464. /* Abort any SDMA transfers */
  465. mpsc_sdma_cmd(pi, 0);
  466. mpsc_sdma_cmd(pi, SDMA_SDCM_AR | SDMA_SDCM_AT);
  467. /* Clear the SDMA current and first TX and RX pointers */
  468. mpsc_sdma_set_tx_ring(pi, NULL);
  469. mpsc_sdma_set_rx_ring(pi, NULL);
  470. /* Disable interrupts */
  471. mpsc_sdma_intr_mask(pi, 0xf);
  472. mpsc_sdma_intr_ack(pi);
  473. }
  474. /*
  475. ******************************************************************************
  476. *
  477. * Multi-Protocol Serial Controller Routines (MPSC)
  478. *
  479. ******************************************************************************
  480. */
  481. static void mpsc_hw_init(struct mpsc_port_info *pi)
  482. {
  483. u32 v;
  484. pr_debug("mpsc_hw_init[%d]: Initializing hardware\n", pi->port.line);
  485. /* Set up clock routing */
  486. if (pi->mirror_regs) {
  487. v = pi->shared_regs->MPSC_MRR_m;
  488. v &= ~0x1c7;
  489. pi->shared_regs->MPSC_MRR_m = v;
  490. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  491. v = pi->shared_regs->MPSC_RCRR_m;
  492. v = (v & ~0xf0f) | 0x100;
  493. pi->shared_regs->MPSC_RCRR_m = v;
  494. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  495. v = pi->shared_regs->MPSC_TCRR_m;
  496. v = (v & ~0xf0f) | 0x100;
  497. pi->shared_regs->MPSC_TCRR_m = v;
  498. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  499. } else {
  500. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  501. v &= ~0x1c7;
  502. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  503. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  504. v = (v & ~0xf0f) | 0x100;
  505. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  506. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  507. v = (v & ~0xf0f) | 0x100;
  508. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  509. }
  510. /* Put MPSC in UART mode & enabel Tx/Rx egines */
  511. writel(0x000004c4, pi->mpsc_base + MPSC_MMCRL);
  512. /* No preamble, 16x divider, low-latency, */
  513. writel(0x04400400, pi->mpsc_base + MPSC_MMCRH);
  514. mpsc_set_baudrate(pi, pi->default_baud);
  515. if (pi->mirror_regs) {
  516. pi->MPSC_CHR_1_m = 0;
  517. pi->MPSC_CHR_2_m = 0;
  518. }
  519. writel(0, pi->mpsc_base + MPSC_CHR_1);
  520. writel(0, pi->mpsc_base + MPSC_CHR_2);
  521. writel(pi->mpsc_max_idle, pi->mpsc_base + MPSC_CHR_3);
  522. writel(0, pi->mpsc_base + MPSC_CHR_4);
  523. writel(0, pi->mpsc_base + MPSC_CHR_5);
  524. writel(0, pi->mpsc_base + MPSC_CHR_6);
  525. writel(0, pi->mpsc_base + MPSC_CHR_7);
  526. writel(0, pi->mpsc_base + MPSC_CHR_8);
  527. writel(0, pi->mpsc_base + MPSC_CHR_9);
  528. writel(0, pi->mpsc_base + MPSC_CHR_10);
  529. }
  530. static void mpsc_enter_hunt(struct mpsc_port_info *pi)
  531. {
  532. pr_debug("mpsc_enter_hunt[%d]: Hunting...\n", pi->port.line);
  533. if (pi->mirror_regs) {
  534. writel(pi->MPSC_CHR_2_m | MPSC_CHR_2_EH,
  535. pi->mpsc_base + MPSC_CHR_2);
  536. /* Erratum prevents reading CHR_2 so just delay for a while */
  537. udelay(100);
  538. } else {
  539. writel(readl(pi->mpsc_base + MPSC_CHR_2) | MPSC_CHR_2_EH,
  540. pi->mpsc_base + MPSC_CHR_2);
  541. while (readl(pi->mpsc_base + MPSC_CHR_2) & MPSC_CHR_2_EH)
  542. udelay(10);
  543. }
  544. }
  545. static void mpsc_freeze(struct mpsc_port_info *pi)
  546. {
  547. u32 v;
  548. pr_debug("mpsc_freeze[%d]: Freezing\n", pi->port.line);
  549. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  550. readl(pi->mpsc_base + MPSC_MPCR);
  551. v |= MPSC_MPCR_FRZ;
  552. if (pi->mirror_regs)
  553. pi->MPSC_MPCR_m = v;
  554. writel(v, pi->mpsc_base + MPSC_MPCR);
  555. }
  556. static void mpsc_unfreeze(struct mpsc_port_info *pi)
  557. {
  558. u32 v;
  559. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  560. readl(pi->mpsc_base + MPSC_MPCR);
  561. v &= ~MPSC_MPCR_FRZ;
  562. if (pi->mirror_regs)
  563. pi->MPSC_MPCR_m = v;
  564. writel(v, pi->mpsc_base + MPSC_MPCR);
  565. pr_debug("mpsc_unfreeze[%d]: Unfrozen\n", pi->port.line);
  566. }
  567. static void mpsc_set_char_length(struct mpsc_port_info *pi, u32 len)
  568. {
  569. u32 v;
  570. pr_debug("mpsc_set_char_length[%d]: char len: %d\n", pi->port.line,len);
  571. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  572. readl(pi->mpsc_base + MPSC_MPCR);
  573. v = (v & ~(0x3 << 12)) | ((len & 0x3) << 12);
  574. if (pi->mirror_regs)
  575. pi->MPSC_MPCR_m = v;
  576. writel(v, pi->mpsc_base + MPSC_MPCR);
  577. }
  578. static void mpsc_set_stop_bit_length(struct mpsc_port_info *pi, u32 len)
  579. {
  580. u32 v;
  581. pr_debug("mpsc_set_stop_bit_length[%d]: stop bits: %d\n",
  582. pi->port.line, len);
  583. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  584. readl(pi->mpsc_base + MPSC_MPCR);
  585. v = (v & ~(1 << 14)) | ((len & 0x1) << 14);
  586. if (pi->mirror_regs)
  587. pi->MPSC_MPCR_m = v;
  588. writel(v, pi->mpsc_base + MPSC_MPCR);
  589. }
  590. static void mpsc_set_parity(struct mpsc_port_info *pi, u32 p)
  591. {
  592. u32 v;
  593. pr_debug("mpsc_set_parity[%d]: parity bits: 0x%x\n", pi->port.line, p);
  594. v = (pi->mirror_regs) ? pi->MPSC_CHR_2_m :
  595. readl(pi->mpsc_base + MPSC_CHR_2);
  596. p &= 0x3;
  597. v = (v & ~0xc000c) | (p << 18) | (p << 2);
  598. if (pi->mirror_regs)
  599. pi->MPSC_CHR_2_m = v;
  600. writel(v, pi->mpsc_base + MPSC_CHR_2);
  601. }
  602. /*
  603. ******************************************************************************
  604. *
  605. * Driver Init Routines
  606. *
  607. ******************************************************************************
  608. */
  609. static void mpsc_init_hw(struct mpsc_port_info *pi)
  610. {
  611. pr_debug("mpsc_init_hw[%d]: Initializing\n", pi->port.line);
  612. mpsc_brg_init(pi, pi->brg_clk_src);
  613. mpsc_brg_enable(pi);
  614. mpsc_sdma_init(pi, dma_get_cache_alignment()); /* burst a cacheline */
  615. mpsc_sdma_stop(pi);
  616. mpsc_hw_init(pi);
  617. }
  618. static int mpsc_alloc_ring_mem(struct mpsc_port_info *pi)
  619. {
  620. int rc = 0;
  621. pr_debug("mpsc_alloc_ring_mem[%d]: Allocating ring mem\n",
  622. pi->port.line);
  623. if (!pi->dma_region) {
  624. if (!dma_set_mask(pi->port.dev, 0xffffffff)) {
  625. printk(KERN_ERR "MPSC: Inadequate DMA support\n");
  626. rc = -ENXIO;
  627. } else if ((pi->dma_region = dma_alloc_attrs(pi->port.dev,
  628. MPSC_DMA_ALLOC_SIZE,
  629. &pi->dma_region_p, GFP_KERNEL,
  630. DMA_ATTR_NON_CONSISTENT))
  631. == NULL) {
  632. printk(KERN_ERR "MPSC: Can't alloc Desc region\n");
  633. rc = -ENOMEM;
  634. }
  635. }
  636. return rc;
  637. }
  638. static void mpsc_free_ring_mem(struct mpsc_port_info *pi)
  639. {
  640. pr_debug("mpsc_free_ring_mem[%d]: Freeing ring mem\n", pi->port.line);
  641. if (pi->dma_region) {
  642. dma_free_attrs(pi->port.dev, MPSC_DMA_ALLOC_SIZE,
  643. pi->dma_region, pi->dma_region_p,
  644. DMA_ATTR_NON_CONSISTENT);
  645. pi->dma_region = NULL;
  646. pi->dma_region_p = (dma_addr_t)NULL;
  647. }
  648. }
  649. static void mpsc_init_rings(struct mpsc_port_info *pi)
  650. {
  651. struct mpsc_rx_desc *rxre;
  652. struct mpsc_tx_desc *txre;
  653. dma_addr_t dp, dp_p;
  654. u8 *bp, *bp_p;
  655. int i;
  656. pr_debug("mpsc_init_rings[%d]: Initializing rings\n", pi->port.line);
  657. BUG_ON(pi->dma_region == NULL);
  658. memset(pi->dma_region, 0, MPSC_DMA_ALLOC_SIZE);
  659. /*
  660. * Descriptors & buffers are multiples of cacheline size and must be
  661. * cacheline aligned.
  662. */
  663. dp = ALIGN((u32)pi->dma_region, dma_get_cache_alignment());
  664. dp_p = ALIGN((u32)pi->dma_region_p, dma_get_cache_alignment());
  665. /*
  666. * Partition dma region into rx ring descriptor, rx buffers,
  667. * tx ring descriptors, and tx buffers.
  668. */
  669. pi->rxr = dp;
  670. pi->rxr_p = dp_p;
  671. dp += MPSC_RXR_SIZE;
  672. dp_p += MPSC_RXR_SIZE;
  673. pi->rxb = (u8 *)dp;
  674. pi->rxb_p = (u8 *)dp_p;
  675. dp += MPSC_RXB_SIZE;
  676. dp_p += MPSC_RXB_SIZE;
  677. pi->rxr_posn = 0;
  678. pi->txr = dp;
  679. pi->txr_p = dp_p;
  680. dp += MPSC_TXR_SIZE;
  681. dp_p += MPSC_TXR_SIZE;
  682. pi->txb = (u8 *)dp;
  683. pi->txb_p = (u8 *)dp_p;
  684. pi->txr_head = 0;
  685. pi->txr_tail = 0;
  686. /* Init rx ring descriptors */
  687. dp = pi->rxr;
  688. dp_p = pi->rxr_p;
  689. bp = pi->rxb;
  690. bp_p = pi->rxb_p;
  691. for (i = 0; i < MPSC_RXR_ENTRIES; i++) {
  692. rxre = (struct mpsc_rx_desc *)dp;
  693. rxre->bufsize = cpu_to_be16(MPSC_RXBE_SIZE);
  694. rxre->bytecnt = cpu_to_be16(0);
  695. rxre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O
  696. | SDMA_DESC_CMDSTAT_EI | SDMA_DESC_CMDSTAT_F
  697. | SDMA_DESC_CMDSTAT_L);
  698. rxre->link = cpu_to_be32(dp_p + MPSC_RXRE_SIZE);
  699. rxre->buf_ptr = cpu_to_be32(bp_p);
  700. dp += MPSC_RXRE_SIZE;
  701. dp_p += MPSC_RXRE_SIZE;
  702. bp += MPSC_RXBE_SIZE;
  703. bp_p += MPSC_RXBE_SIZE;
  704. }
  705. rxre->link = cpu_to_be32(pi->rxr_p); /* Wrap last back to first */
  706. /* Init tx ring descriptors */
  707. dp = pi->txr;
  708. dp_p = pi->txr_p;
  709. bp = pi->txb;
  710. bp_p = pi->txb_p;
  711. for (i = 0; i < MPSC_TXR_ENTRIES; i++) {
  712. txre = (struct mpsc_tx_desc *)dp;
  713. txre->link = cpu_to_be32(dp_p + MPSC_TXRE_SIZE);
  714. txre->buf_ptr = cpu_to_be32(bp_p);
  715. dp += MPSC_TXRE_SIZE;
  716. dp_p += MPSC_TXRE_SIZE;
  717. bp += MPSC_TXBE_SIZE;
  718. bp_p += MPSC_TXBE_SIZE;
  719. }
  720. txre->link = cpu_to_be32(pi->txr_p); /* Wrap last back to first */
  721. dma_cache_sync(pi->port.dev, (void *)pi->dma_region,
  722. MPSC_DMA_ALLOC_SIZE, DMA_BIDIRECTIONAL);
  723. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  724. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  725. flush_dcache_range((ulong)pi->dma_region,
  726. (ulong)pi->dma_region
  727. + MPSC_DMA_ALLOC_SIZE);
  728. #endif
  729. return;
  730. }
  731. static void mpsc_uninit_rings(struct mpsc_port_info *pi)
  732. {
  733. pr_debug("mpsc_uninit_rings[%d]: Uninitializing rings\n",pi->port.line);
  734. BUG_ON(pi->dma_region == NULL);
  735. pi->rxr = 0;
  736. pi->rxr_p = 0;
  737. pi->rxb = NULL;
  738. pi->rxb_p = NULL;
  739. pi->rxr_posn = 0;
  740. pi->txr = 0;
  741. pi->txr_p = 0;
  742. pi->txb = NULL;
  743. pi->txb_p = NULL;
  744. pi->txr_head = 0;
  745. pi->txr_tail = 0;
  746. }
  747. static int mpsc_make_ready(struct mpsc_port_info *pi)
  748. {
  749. int rc;
  750. pr_debug("mpsc_make_ready[%d]: Making cltr ready\n", pi->port.line);
  751. if (!pi->ready) {
  752. mpsc_init_hw(pi);
  753. rc = mpsc_alloc_ring_mem(pi);
  754. if (rc)
  755. return rc;
  756. mpsc_init_rings(pi);
  757. pi->ready = 1;
  758. }
  759. return 0;
  760. }
  761. #ifdef CONFIG_CONSOLE_POLL
  762. static int serial_polled;
  763. #endif
  764. /*
  765. ******************************************************************************
  766. *
  767. * Interrupt Handling Routines
  768. *
  769. ******************************************************************************
  770. */
  771. static int mpsc_rx_intr(struct mpsc_port_info *pi, unsigned long *flags)
  772. {
  773. struct mpsc_rx_desc *rxre;
  774. struct tty_port *port = &pi->port.state->port;
  775. u32 cmdstat, bytes_in, i;
  776. int rc = 0;
  777. u8 *bp;
  778. char flag = TTY_NORMAL;
  779. pr_debug("mpsc_rx_intr[%d]: Handling Rx intr\n", pi->port.line);
  780. rxre = (struct mpsc_rx_desc *)(pi->rxr + (pi->rxr_posn*MPSC_RXRE_SIZE));
  781. dma_cache_sync(pi->port.dev, (void *)rxre, MPSC_RXRE_SIZE,
  782. DMA_FROM_DEVICE);
  783. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  784. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  785. invalidate_dcache_range((ulong)rxre,
  786. (ulong)rxre + MPSC_RXRE_SIZE);
  787. #endif
  788. /*
  789. * Loop through Rx descriptors handling ones that have been completed.
  790. */
  791. while (!((cmdstat = be32_to_cpu(rxre->cmdstat))
  792. & SDMA_DESC_CMDSTAT_O)) {
  793. bytes_in = be16_to_cpu(rxre->bytecnt);
  794. #ifdef CONFIG_CONSOLE_POLL
  795. if (unlikely(serial_polled)) {
  796. serial_polled = 0;
  797. return 0;
  798. }
  799. #endif
  800. /* Following use of tty struct directly is deprecated */
  801. if (tty_buffer_request_room(port, bytes_in) < bytes_in) {
  802. if (port->low_latency) {
  803. spin_unlock_irqrestore(&pi->port.lock, *flags);
  804. tty_flip_buffer_push(port);
  805. spin_lock_irqsave(&pi->port.lock, *flags);
  806. }
  807. /*
  808. * If this failed then we will throw away the bytes
  809. * but must do so to clear interrupts.
  810. */
  811. }
  812. bp = pi->rxb + (pi->rxr_posn * MPSC_RXBE_SIZE);
  813. dma_cache_sync(pi->port.dev, (void *)bp, MPSC_RXBE_SIZE,
  814. DMA_FROM_DEVICE);
  815. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  816. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  817. invalidate_dcache_range((ulong)bp,
  818. (ulong)bp + MPSC_RXBE_SIZE);
  819. #endif
  820. /*
  821. * Other than for parity error, the manual provides little
  822. * info on what data will be in a frame flagged by any of
  823. * these errors. For parity error, it is the last byte in
  824. * the buffer that had the error. As for the rest, I guess
  825. * we'll assume there is no data in the buffer.
  826. * If there is...it gets lost.
  827. */
  828. if (unlikely(cmdstat & (SDMA_DESC_CMDSTAT_BR
  829. | SDMA_DESC_CMDSTAT_FR
  830. | SDMA_DESC_CMDSTAT_OR))) {
  831. pi->port.icount.rx++;
  832. if (cmdstat & SDMA_DESC_CMDSTAT_BR) { /* Break */
  833. pi->port.icount.brk++;
  834. if (uart_handle_break(&pi->port))
  835. goto next_frame;
  836. } else if (cmdstat & SDMA_DESC_CMDSTAT_FR) {
  837. pi->port.icount.frame++;
  838. } else if (cmdstat & SDMA_DESC_CMDSTAT_OR) {
  839. pi->port.icount.overrun++;
  840. }
  841. cmdstat &= pi->port.read_status_mask;
  842. if (cmdstat & SDMA_DESC_CMDSTAT_BR)
  843. flag = TTY_BREAK;
  844. else if (cmdstat & SDMA_DESC_CMDSTAT_FR)
  845. flag = TTY_FRAME;
  846. else if (cmdstat & SDMA_DESC_CMDSTAT_OR)
  847. flag = TTY_OVERRUN;
  848. else if (cmdstat & SDMA_DESC_CMDSTAT_PE)
  849. flag = TTY_PARITY;
  850. }
  851. if (uart_handle_sysrq_char(&pi->port, *bp)) {
  852. bp++;
  853. bytes_in--;
  854. #ifdef CONFIG_CONSOLE_POLL
  855. if (unlikely(serial_polled)) {
  856. serial_polled = 0;
  857. return 0;
  858. }
  859. #endif
  860. goto next_frame;
  861. }
  862. if ((unlikely(cmdstat & (SDMA_DESC_CMDSTAT_BR
  863. | SDMA_DESC_CMDSTAT_FR
  864. | SDMA_DESC_CMDSTAT_OR)))
  865. && !(cmdstat & pi->port.ignore_status_mask)) {
  866. tty_insert_flip_char(port, *bp, flag);
  867. } else {
  868. for (i=0; i<bytes_in; i++)
  869. tty_insert_flip_char(port, *bp++, TTY_NORMAL);
  870. pi->port.icount.rx += bytes_in;
  871. }
  872. next_frame:
  873. rxre->bytecnt = cpu_to_be16(0);
  874. wmb();
  875. rxre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O
  876. | SDMA_DESC_CMDSTAT_EI | SDMA_DESC_CMDSTAT_F
  877. | SDMA_DESC_CMDSTAT_L);
  878. wmb();
  879. dma_cache_sync(pi->port.dev, (void *)rxre, MPSC_RXRE_SIZE,
  880. DMA_BIDIRECTIONAL);
  881. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  882. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  883. flush_dcache_range((ulong)rxre,
  884. (ulong)rxre + MPSC_RXRE_SIZE);
  885. #endif
  886. /* Advance to next descriptor */
  887. pi->rxr_posn = (pi->rxr_posn + 1) & (MPSC_RXR_ENTRIES - 1);
  888. rxre = (struct mpsc_rx_desc *)
  889. (pi->rxr + (pi->rxr_posn * MPSC_RXRE_SIZE));
  890. dma_cache_sync(pi->port.dev, (void *)rxre, MPSC_RXRE_SIZE,
  891. DMA_FROM_DEVICE);
  892. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  893. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  894. invalidate_dcache_range((ulong)rxre,
  895. (ulong)rxre + MPSC_RXRE_SIZE);
  896. #endif
  897. rc = 1;
  898. }
  899. /* Restart rx engine, if its stopped */
  900. if ((readl(pi->sdma_base + SDMA_SDCM) & SDMA_SDCM_ERD) == 0)
  901. mpsc_start_rx(pi);
  902. spin_unlock_irqrestore(&pi->port.lock, *flags);
  903. tty_flip_buffer_push(port);
  904. spin_lock_irqsave(&pi->port.lock, *flags);
  905. return rc;
  906. }
  907. static void mpsc_setup_tx_desc(struct mpsc_port_info *pi, u32 count, u32 intr)
  908. {
  909. struct mpsc_tx_desc *txre;
  910. txre = (struct mpsc_tx_desc *)(pi->txr
  911. + (pi->txr_head * MPSC_TXRE_SIZE));
  912. txre->bytecnt = cpu_to_be16(count);
  913. txre->shadow = txre->bytecnt;
  914. wmb(); /* ensure cmdstat is last field updated */
  915. txre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O | SDMA_DESC_CMDSTAT_F
  916. | SDMA_DESC_CMDSTAT_L
  917. | ((intr) ? SDMA_DESC_CMDSTAT_EI : 0));
  918. wmb();
  919. dma_cache_sync(pi->port.dev, (void *)txre, MPSC_TXRE_SIZE,
  920. DMA_BIDIRECTIONAL);
  921. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  922. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  923. flush_dcache_range((ulong)txre,
  924. (ulong)txre + MPSC_TXRE_SIZE);
  925. #endif
  926. }
  927. static void mpsc_copy_tx_data(struct mpsc_port_info *pi)
  928. {
  929. struct circ_buf *xmit = &pi->port.state->xmit;
  930. u8 *bp;
  931. u32 i;
  932. /* Make sure the desc ring isn't full */
  933. while (CIRC_CNT(pi->txr_head, pi->txr_tail, MPSC_TXR_ENTRIES)
  934. < (MPSC_TXR_ENTRIES - 1)) {
  935. if (pi->port.x_char) {
  936. /*
  937. * Ideally, we should use the TCS field in
  938. * CHR_1 to put the x_char out immediately but
  939. * errata prevents us from being able to read
  940. * CHR_2 to know that its safe to write to
  941. * CHR_1. Instead, just put it in-band with
  942. * all the other Tx data.
  943. */
  944. bp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  945. *bp = pi->port.x_char;
  946. pi->port.x_char = 0;
  947. i = 1;
  948. } else if (!uart_circ_empty(xmit)
  949. && !uart_tx_stopped(&pi->port)) {
  950. i = min((u32)MPSC_TXBE_SIZE,
  951. (u32)uart_circ_chars_pending(xmit));
  952. i = min(i, (u32)CIRC_CNT_TO_END(xmit->head, xmit->tail,
  953. UART_XMIT_SIZE));
  954. bp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  955. memcpy(bp, &xmit->buf[xmit->tail], i);
  956. xmit->tail = (xmit->tail + i) & (UART_XMIT_SIZE - 1);
  957. if (uart_circ_chars_pending(xmit) < WAKEUP_CHARS)
  958. uart_write_wakeup(&pi->port);
  959. } else { /* All tx data copied into ring bufs */
  960. return;
  961. }
  962. dma_cache_sync(pi->port.dev, (void *)bp, MPSC_TXBE_SIZE,
  963. DMA_BIDIRECTIONAL);
  964. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  965. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  966. flush_dcache_range((ulong)bp,
  967. (ulong)bp + MPSC_TXBE_SIZE);
  968. #endif
  969. mpsc_setup_tx_desc(pi, i, 1);
  970. /* Advance to next descriptor */
  971. pi->txr_head = (pi->txr_head + 1) & (MPSC_TXR_ENTRIES - 1);
  972. }
  973. }
  974. static int mpsc_tx_intr(struct mpsc_port_info *pi)
  975. {
  976. struct mpsc_tx_desc *txre;
  977. int rc = 0;
  978. unsigned long iflags;
  979. spin_lock_irqsave(&pi->tx_lock, iflags);
  980. if (!mpsc_sdma_tx_active(pi)) {
  981. txre = (struct mpsc_tx_desc *)(pi->txr
  982. + (pi->txr_tail * MPSC_TXRE_SIZE));
  983. dma_cache_sync(pi->port.dev, (void *)txre, MPSC_TXRE_SIZE,
  984. DMA_FROM_DEVICE);
  985. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  986. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  987. invalidate_dcache_range((ulong)txre,
  988. (ulong)txre + MPSC_TXRE_SIZE);
  989. #endif
  990. while (!(be32_to_cpu(txre->cmdstat) & SDMA_DESC_CMDSTAT_O)) {
  991. rc = 1;
  992. pi->port.icount.tx += be16_to_cpu(txre->bytecnt);
  993. pi->txr_tail = (pi->txr_tail+1) & (MPSC_TXR_ENTRIES-1);
  994. /* If no more data to tx, fall out of loop */
  995. if (pi->txr_head == pi->txr_tail)
  996. break;
  997. txre = (struct mpsc_tx_desc *)(pi->txr
  998. + (pi->txr_tail * MPSC_TXRE_SIZE));
  999. dma_cache_sync(pi->port.dev, (void *)txre,
  1000. MPSC_TXRE_SIZE, DMA_FROM_DEVICE);
  1001. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  1002. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1003. invalidate_dcache_range((ulong)txre,
  1004. (ulong)txre + MPSC_TXRE_SIZE);
  1005. #endif
  1006. }
  1007. mpsc_copy_tx_data(pi);
  1008. mpsc_sdma_start_tx(pi); /* start next desc if ready */
  1009. }
  1010. spin_unlock_irqrestore(&pi->tx_lock, iflags);
  1011. return rc;
  1012. }
  1013. /*
  1014. * This is the driver's interrupt handler. To avoid a race, we first clear
  1015. * the interrupt, then handle any completed Rx/Tx descriptors. When done
  1016. * handling those descriptors, we restart the Rx/Tx engines if they're stopped.
  1017. */
  1018. static irqreturn_t mpsc_sdma_intr(int irq, void *dev_id)
  1019. {
  1020. struct mpsc_port_info *pi = dev_id;
  1021. ulong iflags;
  1022. int rc = IRQ_NONE;
  1023. pr_debug("mpsc_sdma_intr[%d]: SDMA Interrupt Received\n",pi->port.line);
  1024. spin_lock_irqsave(&pi->port.lock, iflags);
  1025. mpsc_sdma_intr_ack(pi);
  1026. if (mpsc_rx_intr(pi, &iflags))
  1027. rc = IRQ_HANDLED;
  1028. if (mpsc_tx_intr(pi))
  1029. rc = IRQ_HANDLED;
  1030. spin_unlock_irqrestore(&pi->port.lock, iflags);
  1031. pr_debug("mpsc_sdma_intr[%d]: SDMA Interrupt Handled\n", pi->port.line);
  1032. return rc;
  1033. }
  1034. /*
  1035. ******************************************************************************
  1036. *
  1037. * serial_core.c Interface routines
  1038. *
  1039. ******************************************************************************
  1040. */
  1041. static uint mpsc_tx_empty(struct uart_port *port)
  1042. {
  1043. struct mpsc_port_info *pi =
  1044. container_of(port, struct mpsc_port_info, port);
  1045. ulong iflags;
  1046. uint rc;
  1047. spin_lock_irqsave(&pi->port.lock, iflags);
  1048. rc = mpsc_sdma_tx_active(pi) ? 0 : TIOCSER_TEMT;
  1049. spin_unlock_irqrestore(&pi->port.lock, iflags);
  1050. return rc;
  1051. }
  1052. static void mpsc_set_mctrl(struct uart_port *port, uint mctrl)
  1053. {
  1054. /* Have no way to set modem control lines AFAICT */
  1055. }
  1056. static uint mpsc_get_mctrl(struct uart_port *port)
  1057. {
  1058. struct mpsc_port_info *pi =
  1059. container_of(port, struct mpsc_port_info, port);
  1060. u32 mflags, status;
  1061. status = (pi->mirror_regs) ? pi->MPSC_CHR_10_m
  1062. : readl(pi->mpsc_base + MPSC_CHR_10);
  1063. mflags = 0;
  1064. if (status & 0x1)
  1065. mflags |= TIOCM_CTS;
  1066. if (status & 0x2)
  1067. mflags |= TIOCM_CAR;
  1068. return mflags | TIOCM_DSR; /* No way to tell if DSR asserted */
  1069. }
  1070. static void mpsc_stop_tx(struct uart_port *port)
  1071. {
  1072. struct mpsc_port_info *pi =
  1073. container_of(port, struct mpsc_port_info, port);
  1074. pr_debug("mpsc_stop_tx[%d]\n", port->line);
  1075. mpsc_freeze(pi);
  1076. }
  1077. static void mpsc_start_tx(struct uart_port *port)
  1078. {
  1079. struct mpsc_port_info *pi =
  1080. container_of(port, struct mpsc_port_info, port);
  1081. unsigned long iflags;
  1082. spin_lock_irqsave(&pi->tx_lock, iflags);
  1083. mpsc_unfreeze(pi);
  1084. mpsc_copy_tx_data(pi);
  1085. mpsc_sdma_start_tx(pi);
  1086. spin_unlock_irqrestore(&pi->tx_lock, iflags);
  1087. pr_debug("mpsc_start_tx[%d]\n", port->line);
  1088. }
  1089. static void mpsc_start_rx(struct mpsc_port_info *pi)
  1090. {
  1091. pr_debug("mpsc_start_rx[%d]: Starting...\n", pi->port.line);
  1092. if (pi->rcv_data) {
  1093. mpsc_enter_hunt(pi);
  1094. mpsc_sdma_cmd(pi, SDMA_SDCM_ERD);
  1095. }
  1096. }
  1097. static void mpsc_stop_rx(struct uart_port *port)
  1098. {
  1099. struct mpsc_port_info *pi =
  1100. container_of(port, struct mpsc_port_info, port);
  1101. pr_debug("mpsc_stop_rx[%d]: Stopping...\n", port->line);
  1102. if (pi->mirror_regs) {
  1103. writel(pi->MPSC_CHR_2_m | MPSC_CHR_2_RA,
  1104. pi->mpsc_base + MPSC_CHR_2);
  1105. /* Erratum prevents reading CHR_2 so just delay for a while */
  1106. udelay(100);
  1107. } else {
  1108. writel(readl(pi->mpsc_base + MPSC_CHR_2) | MPSC_CHR_2_RA,
  1109. pi->mpsc_base + MPSC_CHR_2);
  1110. while (readl(pi->mpsc_base + MPSC_CHR_2) & MPSC_CHR_2_RA)
  1111. udelay(10);
  1112. }
  1113. mpsc_sdma_cmd(pi, SDMA_SDCM_AR);
  1114. }
  1115. static void mpsc_break_ctl(struct uart_port *port, int ctl)
  1116. {
  1117. struct mpsc_port_info *pi =
  1118. container_of(port, struct mpsc_port_info, port);
  1119. ulong flags;
  1120. u32 v;
  1121. v = ctl ? 0x00ff0000 : 0;
  1122. spin_lock_irqsave(&pi->port.lock, flags);
  1123. if (pi->mirror_regs)
  1124. pi->MPSC_CHR_1_m = v;
  1125. writel(v, pi->mpsc_base + MPSC_CHR_1);
  1126. spin_unlock_irqrestore(&pi->port.lock, flags);
  1127. }
  1128. static int mpsc_startup(struct uart_port *port)
  1129. {
  1130. struct mpsc_port_info *pi =
  1131. container_of(port, struct mpsc_port_info, port);
  1132. u32 flag = 0;
  1133. int rc;
  1134. pr_debug("mpsc_startup[%d]: Starting up MPSC, irq: %d\n",
  1135. port->line, pi->port.irq);
  1136. if ((rc = mpsc_make_ready(pi)) == 0) {
  1137. /* Setup IRQ handler */
  1138. mpsc_sdma_intr_ack(pi);
  1139. /* If irq's are shared, need to set flag */
  1140. if (mpsc_ports[0].port.irq == mpsc_ports[1].port.irq)
  1141. flag = IRQF_SHARED;
  1142. if (request_irq(pi->port.irq, mpsc_sdma_intr, flag,
  1143. "mpsc-sdma", pi))
  1144. printk(KERN_ERR "MPSC: Can't get SDMA IRQ %d\n",
  1145. pi->port.irq);
  1146. mpsc_sdma_intr_unmask(pi, 0xf);
  1147. mpsc_sdma_set_rx_ring(pi, (struct mpsc_rx_desc *)(pi->rxr_p
  1148. + (pi->rxr_posn * MPSC_RXRE_SIZE)));
  1149. }
  1150. return rc;
  1151. }
  1152. static void mpsc_shutdown(struct uart_port *port)
  1153. {
  1154. struct mpsc_port_info *pi =
  1155. container_of(port, struct mpsc_port_info, port);
  1156. pr_debug("mpsc_shutdown[%d]: Shutting down MPSC\n", port->line);
  1157. mpsc_sdma_stop(pi);
  1158. free_irq(pi->port.irq, pi);
  1159. }
  1160. static void mpsc_set_termios(struct uart_port *port, struct ktermios *termios,
  1161. struct ktermios *old)
  1162. {
  1163. struct mpsc_port_info *pi =
  1164. container_of(port, struct mpsc_port_info, port);
  1165. u32 baud;
  1166. ulong flags;
  1167. u32 chr_bits, stop_bits, par;
  1168. switch (termios->c_cflag & CSIZE) {
  1169. case CS5:
  1170. chr_bits = MPSC_MPCR_CL_5;
  1171. break;
  1172. case CS6:
  1173. chr_bits = MPSC_MPCR_CL_6;
  1174. break;
  1175. case CS7:
  1176. chr_bits = MPSC_MPCR_CL_7;
  1177. break;
  1178. case CS8:
  1179. default:
  1180. chr_bits = MPSC_MPCR_CL_8;
  1181. break;
  1182. }
  1183. if (termios->c_cflag & CSTOPB)
  1184. stop_bits = MPSC_MPCR_SBL_2;
  1185. else
  1186. stop_bits = MPSC_MPCR_SBL_1;
  1187. par = MPSC_CHR_2_PAR_EVEN;
  1188. if (termios->c_cflag & PARENB)
  1189. if (termios->c_cflag & PARODD)
  1190. par = MPSC_CHR_2_PAR_ODD;
  1191. #ifdef CMSPAR
  1192. if (termios->c_cflag & CMSPAR) {
  1193. if (termios->c_cflag & PARODD)
  1194. par = MPSC_CHR_2_PAR_MARK;
  1195. else
  1196. par = MPSC_CHR_2_PAR_SPACE;
  1197. }
  1198. #endif
  1199. baud = uart_get_baud_rate(port, termios, old, 0, port->uartclk);
  1200. spin_lock_irqsave(&pi->port.lock, flags);
  1201. uart_update_timeout(port, termios->c_cflag, baud);
  1202. mpsc_set_char_length(pi, chr_bits);
  1203. mpsc_set_stop_bit_length(pi, stop_bits);
  1204. mpsc_set_parity(pi, par);
  1205. mpsc_set_baudrate(pi, baud);
  1206. /* Characters/events to read */
  1207. pi->port.read_status_mask = SDMA_DESC_CMDSTAT_OR;
  1208. if (termios->c_iflag & INPCK)
  1209. pi->port.read_status_mask |= SDMA_DESC_CMDSTAT_PE
  1210. | SDMA_DESC_CMDSTAT_FR;
  1211. if (termios->c_iflag & (IGNBRK | BRKINT | PARMRK))
  1212. pi->port.read_status_mask |= SDMA_DESC_CMDSTAT_BR;
  1213. /* Characters/events to ignore */
  1214. pi->port.ignore_status_mask = 0;
  1215. if (termios->c_iflag & IGNPAR)
  1216. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_PE
  1217. | SDMA_DESC_CMDSTAT_FR;
  1218. if (termios->c_iflag & IGNBRK) {
  1219. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_BR;
  1220. if (termios->c_iflag & IGNPAR)
  1221. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_OR;
  1222. }
  1223. if ((termios->c_cflag & CREAD)) {
  1224. if (!pi->rcv_data) {
  1225. pi->rcv_data = 1;
  1226. mpsc_start_rx(pi);
  1227. }
  1228. } else if (pi->rcv_data) {
  1229. mpsc_stop_rx(port);
  1230. pi->rcv_data = 0;
  1231. }
  1232. spin_unlock_irqrestore(&pi->port.lock, flags);
  1233. }
  1234. static const char *mpsc_type(struct uart_port *port)
  1235. {
  1236. pr_debug("mpsc_type[%d]: port type: %s\n", port->line,MPSC_DRIVER_NAME);
  1237. return MPSC_DRIVER_NAME;
  1238. }
  1239. static int mpsc_request_port(struct uart_port *port)
  1240. {
  1241. /* Should make chip/platform specific call */
  1242. return 0;
  1243. }
  1244. static void mpsc_release_port(struct uart_port *port)
  1245. {
  1246. struct mpsc_port_info *pi =
  1247. container_of(port, struct mpsc_port_info, port);
  1248. if (pi->ready) {
  1249. mpsc_uninit_rings(pi);
  1250. mpsc_free_ring_mem(pi);
  1251. pi->ready = 0;
  1252. }
  1253. }
  1254. static void mpsc_config_port(struct uart_port *port, int flags)
  1255. {
  1256. }
  1257. static int mpsc_verify_port(struct uart_port *port, struct serial_struct *ser)
  1258. {
  1259. struct mpsc_port_info *pi =
  1260. container_of(port, struct mpsc_port_info, port);
  1261. int rc = 0;
  1262. pr_debug("mpsc_verify_port[%d]: Verifying port data\n", pi->port.line);
  1263. if (ser->type != PORT_UNKNOWN && ser->type != PORT_MPSC)
  1264. rc = -EINVAL;
  1265. else if (pi->port.irq != ser->irq)
  1266. rc = -EINVAL;
  1267. else if (ser->io_type != SERIAL_IO_MEM)
  1268. rc = -EINVAL;
  1269. else if (pi->port.uartclk / 16 != ser->baud_base) /* Not sure */
  1270. rc = -EINVAL;
  1271. else if ((void *)pi->port.mapbase != ser->iomem_base)
  1272. rc = -EINVAL;
  1273. else if (pi->port.iobase != ser->port)
  1274. rc = -EINVAL;
  1275. else if (ser->hub6 != 0)
  1276. rc = -EINVAL;
  1277. return rc;
  1278. }
  1279. #ifdef CONFIG_CONSOLE_POLL
  1280. /* Serial polling routines for writing and reading from the uart while
  1281. * in an interrupt or debug context.
  1282. */
  1283. static char poll_buf[2048];
  1284. static int poll_ptr;
  1285. static int poll_cnt;
  1286. static void mpsc_put_poll_char(struct uart_port *port,
  1287. unsigned char c);
  1288. static int mpsc_get_poll_char(struct uart_port *port)
  1289. {
  1290. struct mpsc_port_info *pi =
  1291. container_of(port, struct mpsc_port_info, port);
  1292. struct mpsc_rx_desc *rxre;
  1293. u32 cmdstat, bytes_in, i;
  1294. u8 *bp;
  1295. if (!serial_polled)
  1296. serial_polled = 1;
  1297. pr_debug("mpsc_rx_intr[%d]: Handling Rx intr\n", pi->port.line);
  1298. if (poll_cnt) {
  1299. poll_cnt--;
  1300. return poll_buf[poll_ptr++];
  1301. }
  1302. poll_ptr = 0;
  1303. poll_cnt = 0;
  1304. while (poll_cnt == 0) {
  1305. rxre = (struct mpsc_rx_desc *)(pi->rxr +
  1306. (pi->rxr_posn*MPSC_RXRE_SIZE));
  1307. dma_cache_sync(pi->port.dev, (void *)rxre,
  1308. MPSC_RXRE_SIZE, DMA_FROM_DEVICE);
  1309. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  1310. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1311. invalidate_dcache_range((ulong)rxre,
  1312. (ulong)rxre + MPSC_RXRE_SIZE);
  1313. #endif
  1314. /*
  1315. * Loop through Rx descriptors handling ones that have
  1316. * been completed.
  1317. */
  1318. while (poll_cnt == 0 &&
  1319. !((cmdstat = be32_to_cpu(rxre->cmdstat)) &
  1320. SDMA_DESC_CMDSTAT_O)){
  1321. bytes_in = be16_to_cpu(rxre->bytecnt);
  1322. bp = pi->rxb + (pi->rxr_posn * MPSC_RXBE_SIZE);
  1323. dma_cache_sync(pi->port.dev, (void *) bp,
  1324. MPSC_RXBE_SIZE, DMA_FROM_DEVICE);
  1325. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  1326. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1327. invalidate_dcache_range((ulong)bp,
  1328. (ulong)bp + MPSC_RXBE_SIZE);
  1329. #endif
  1330. if ((unlikely(cmdstat & (SDMA_DESC_CMDSTAT_BR |
  1331. SDMA_DESC_CMDSTAT_FR | SDMA_DESC_CMDSTAT_OR))) &&
  1332. !(cmdstat & pi->port.ignore_status_mask)) {
  1333. poll_buf[poll_cnt] = *bp;
  1334. poll_cnt++;
  1335. } else {
  1336. for (i = 0; i < bytes_in; i++) {
  1337. poll_buf[poll_cnt] = *bp++;
  1338. poll_cnt++;
  1339. }
  1340. pi->port.icount.rx += bytes_in;
  1341. }
  1342. rxre->bytecnt = cpu_to_be16(0);
  1343. wmb();
  1344. rxre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O |
  1345. SDMA_DESC_CMDSTAT_EI |
  1346. SDMA_DESC_CMDSTAT_F |
  1347. SDMA_DESC_CMDSTAT_L);
  1348. wmb();
  1349. dma_cache_sync(pi->port.dev, (void *)rxre,
  1350. MPSC_RXRE_SIZE, DMA_BIDIRECTIONAL);
  1351. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  1352. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1353. flush_dcache_range((ulong)rxre,
  1354. (ulong)rxre + MPSC_RXRE_SIZE);
  1355. #endif
  1356. /* Advance to next descriptor */
  1357. pi->rxr_posn = (pi->rxr_posn + 1) &
  1358. (MPSC_RXR_ENTRIES - 1);
  1359. rxre = (struct mpsc_rx_desc *)(pi->rxr +
  1360. (pi->rxr_posn * MPSC_RXRE_SIZE));
  1361. dma_cache_sync(pi->port.dev, (void *)rxre,
  1362. MPSC_RXRE_SIZE, DMA_FROM_DEVICE);
  1363. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  1364. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1365. invalidate_dcache_range((ulong)rxre,
  1366. (ulong)rxre + MPSC_RXRE_SIZE);
  1367. #endif
  1368. }
  1369. /* Restart rx engine, if its stopped */
  1370. if ((readl(pi->sdma_base + SDMA_SDCM) & SDMA_SDCM_ERD) == 0)
  1371. mpsc_start_rx(pi);
  1372. }
  1373. if (poll_cnt) {
  1374. poll_cnt--;
  1375. return poll_buf[poll_ptr++];
  1376. }
  1377. return 0;
  1378. }
  1379. static void mpsc_put_poll_char(struct uart_port *port,
  1380. unsigned char c)
  1381. {
  1382. struct mpsc_port_info *pi =
  1383. container_of(port, struct mpsc_port_info, port);
  1384. u32 data;
  1385. data = readl(pi->mpsc_base + MPSC_MPCR);
  1386. writeb(c, pi->mpsc_base + MPSC_CHR_1);
  1387. mb();
  1388. data = readl(pi->mpsc_base + MPSC_CHR_2);
  1389. data |= MPSC_CHR_2_TTCS;
  1390. writel(data, pi->mpsc_base + MPSC_CHR_2);
  1391. mb();
  1392. while (readl(pi->mpsc_base + MPSC_CHR_2) & MPSC_CHR_2_TTCS);
  1393. }
  1394. #endif
  1395. static const struct uart_ops mpsc_pops = {
  1396. .tx_empty = mpsc_tx_empty,
  1397. .set_mctrl = mpsc_set_mctrl,
  1398. .get_mctrl = mpsc_get_mctrl,
  1399. .stop_tx = mpsc_stop_tx,
  1400. .start_tx = mpsc_start_tx,
  1401. .stop_rx = mpsc_stop_rx,
  1402. .break_ctl = mpsc_break_ctl,
  1403. .startup = mpsc_startup,
  1404. .shutdown = mpsc_shutdown,
  1405. .set_termios = mpsc_set_termios,
  1406. .type = mpsc_type,
  1407. .release_port = mpsc_release_port,
  1408. .request_port = mpsc_request_port,
  1409. .config_port = mpsc_config_port,
  1410. .verify_port = mpsc_verify_port,
  1411. #ifdef CONFIG_CONSOLE_POLL
  1412. .poll_get_char = mpsc_get_poll_char,
  1413. .poll_put_char = mpsc_put_poll_char,
  1414. #endif
  1415. };
  1416. /*
  1417. ******************************************************************************
  1418. *
  1419. * Console Interface Routines
  1420. *
  1421. ******************************************************************************
  1422. */
  1423. #ifdef CONFIG_SERIAL_MPSC_CONSOLE
  1424. static void mpsc_console_write(struct console *co, const char *s, uint count)
  1425. {
  1426. struct mpsc_port_info *pi = &mpsc_ports[co->index];
  1427. u8 *bp, *dp, add_cr = 0;
  1428. int i;
  1429. unsigned long iflags;
  1430. spin_lock_irqsave(&pi->tx_lock, iflags);
  1431. while (pi->txr_head != pi->txr_tail) {
  1432. while (mpsc_sdma_tx_active(pi))
  1433. udelay(100);
  1434. mpsc_sdma_intr_ack(pi);
  1435. mpsc_tx_intr(pi);
  1436. }
  1437. while (mpsc_sdma_tx_active(pi))
  1438. udelay(100);
  1439. while (count > 0) {
  1440. bp = dp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  1441. for (i = 0; i < MPSC_TXBE_SIZE; i++) {
  1442. if (count == 0)
  1443. break;
  1444. if (add_cr) {
  1445. *(dp++) = '\r';
  1446. add_cr = 0;
  1447. } else {
  1448. *(dp++) = *s;
  1449. if (*(s++) == '\n') { /* add '\r' after '\n' */
  1450. add_cr = 1;
  1451. count++;
  1452. }
  1453. }
  1454. count--;
  1455. }
  1456. dma_cache_sync(pi->port.dev, (void *)bp, MPSC_TXBE_SIZE,
  1457. DMA_BIDIRECTIONAL);
  1458. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  1459. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1460. flush_dcache_range((ulong)bp,
  1461. (ulong)bp + MPSC_TXBE_SIZE);
  1462. #endif
  1463. mpsc_setup_tx_desc(pi, i, 0);
  1464. pi->txr_head = (pi->txr_head + 1) & (MPSC_TXR_ENTRIES - 1);
  1465. mpsc_sdma_start_tx(pi);
  1466. while (mpsc_sdma_tx_active(pi))
  1467. udelay(100);
  1468. pi->txr_tail = (pi->txr_tail + 1) & (MPSC_TXR_ENTRIES - 1);
  1469. }
  1470. spin_unlock_irqrestore(&pi->tx_lock, iflags);
  1471. }
  1472. static int __init mpsc_console_setup(struct console *co, char *options)
  1473. {
  1474. struct mpsc_port_info *pi;
  1475. int baud, bits, parity, flow;
  1476. pr_debug("mpsc_console_setup[%d]: options: %s\n", co->index, options);
  1477. if (co->index >= MPSC_NUM_CTLRS)
  1478. co->index = 0;
  1479. pi = &mpsc_ports[co->index];
  1480. baud = pi->default_baud;
  1481. bits = pi->default_bits;
  1482. parity = pi->default_parity;
  1483. flow = pi->default_flow;
  1484. if (!pi->port.ops)
  1485. return -ENODEV;
  1486. spin_lock_init(&pi->port.lock); /* Temporary fix--copied from 8250.c */
  1487. if (options)
  1488. uart_parse_options(options, &baud, &parity, &bits, &flow);
  1489. return uart_set_options(&pi->port, co, baud, parity, bits, flow);
  1490. }
  1491. static struct console mpsc_console = {
  1492. .name = MPSC_DEV_NAME,
  1493. .write = mpsc_console_write,
  1494. .device = uart_console_device,
  1495. .setup = mpsc_console_setup,
  1496. .flags = CON_PRINTBUFFER,
  1497. .index = -1,
  1498. .data = &mpsc_reg,
  1499. };
  1500. static int __init mpsc_late_console_init(void)
  1501. {
  1502. pr_debug("mpsc_late_console_init: Enter\n");
  1503. if (!(mpsc_console.flags & CON_ENABLED))
  1504. register_console(&mpsc_console);
  1505. return 0;
  1506. }
  1507. late_initcall(mpsc_late_console_init);
  1508. #define MPSC_CONSOLE &mpsc_console
  1509. #else
  1510. #define MPSC_CONSOLE NULL
  1511. #endif
  1512. /*
  1513. ******************************************************************************
  1514. *
  1515. * Dummy Platform Driver to extract & map shared register regions
  1516. *
  1517. ******************************************************************************
  1518. */
  1519. static void mpsc_resource_err(char *s)
  1520. {
  1521. printk(KERN_WARNING "MPSC: Platform device resource error in %s\n", s);
  1522. }
  1523. static int mpsc_shared_map_regs(struct platform_device *pd)
  1524. {
  1525. struct resource *r;
  1526. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1527. MPSC_ROUTING_BASE_ORDER))
  1528. && request_mem_region(r->start,
  1529. MPSC_ROUTING_REG_BLOCK_SIZE,
  1530. "mpsc_routing_regs")) {
  1531. mpsc_shared_regs.mpsc_routing_base = ioremap(r->start,
  1532. MPSC_ROUTING_REG_BLOCK_SIZE);
  1533. mpsc_shared_regs.mpsc_routing_base_p = r->start;
  1534. } else {
  1535. mpsc_resource_err("MPSC routing base");
  1536. return -ENOMEM;
  1537. }
  1538. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1539. MPSC_SDMA_INTR_BASE_ORDER))
  1540. && request_mem_region(r->start,
  1541. MPSC_SDMA_INTR_REG_BLOCK_SIZE,
  1542. "sdma_intr_regs")) {
  1543. mpsc_shared_regs.sdma_intr_base = ioremap(r->start,
  1544. MPSC_SDMA_INTR_REG_BLOCK_SIZE);
  1545. mpsc_shared_regs.sdma_intr_base_p = r->start;
  1546. } else {
  1547. iounmap(mpsc_shared_regs.mpsc_routing_base);
  1548. release_mem_region(mpsc_shared_regs.mpsc_routing_base_p,
  1549. MPSC_ROUTING_REG_BLOCK_SIZE);
  1550. mpsc_resource_err("SDMA intr base");
  1551. return -ENOMEM;
  1552. }
  1553. return 0;
  1554. }
  1555. static void mpsc_shared_unmap_regs(void)
  1556. {
  1557. if (mpsc_shared_regs.mpsc_routing_base) {
  1558. iounmap(mpsc_shared_regs.mpsc_routing_base);
  1559. release_mem_region(mpsc_shared_regs.mpsc_routing_base_p,
  1560. MPSC_ROUTING_REG_BLOCK_SIZE);
  1561. }
  1562. if (mpsc_shared_regs.sdma_intr_base) {
  1563. iounmap(mpsc_shared_regs.sdma_intr_base);
  1564. release_mem_region(mpsc_shared_regs.sdma_intr_base_p,
  1565. MPSC_SDMA_INTR_REG_BLOCK_SIZE);
  1566. }
  1567. mpsc_shared_regs.mpsc_routing_base = NULL;
  1568. mpsc_shared_regs.sdma_intr_base = NULL;
  1569. mpsc_shared_regs.mpsc_routing_base_p = 0;
  1570. mpsc_shared_regs.sdma_intr_base_p = 0;
  1571. }
  1572. static int mpsc_shared_drv_probe(struct platform_device *dev)
  1573. {
  1574. struct mpsc_shared_pdata *pdata;
  1575. int rc;
  1576. if (dev->id != 0)
  1577. return -ENODEV;
  1578. rc = mpsc_shared_map_regs(dev);
  1579. if (rc)
  1580. return rc;
  1581. pdata = dev_get_platdata(&dev->dev);
  1582. mpsc_shared_regs.MPSC_MRR_m = pdata->mrr_val;
  1583. mpsc_shared_regs.MPSC_RCRR_m= pdata->rcrr_val;
  1584. mpsc_shared_regs.MPSC_TCRR_m= pdata->tcrr_val;
  1585. mpsc_shared_regs.SDMA_INTR_CAUSE_m = pdata->intr_cause_val;
  1586. mpsc_shared_regs.SDMA_INTR_MASK_m = pdata->intr_mask_val;
  1587. return 0;
  1588. }
  1589. static int mpsc_shared_drv_remove(struct platform_device *dev)
  1590. {
  1591. if (dev->id != 0)
  1592. return -ENODEV;
  1593. mpsc_shared_unmap_regs();
  1594. mpsc_shared_regs.MPSC_MRR_m = 0;
  1595. mpsc_shared_regs.MPSC_RCRR_m = 0;
  1596. mpsc_shared_regs.MPSC_TCRR_m = 0;
  1597. mpsc_shared_regs.SDMA_INTR_CAUSE_m = 0;
  1598. mpsc_shared_regs.SDMA_INTR_MASK_m = 0;
  1599. return 0;
  1600. }
  1601. static struct platform_driver mpsc_shared_driver = {
  1602. .probe = mpsc_shared_drv_probe,
  1603. .remove = mpsc_shared_drv_remove,
  1604. .driver = {
  1605. .name = MPSC_SHARED_NAME,
  1606. },
  1607. };
  1608. /*
  1609. ******************************************************************************
  1610. *
  1611. * Driver Interface Routines
  1612. *
  1613. ******************************************************************************
  1614. */
  1615. static struct uart_driver mpsc_reg = {
  1616. .owner = THIS_MODULE,
  1617. .driver_name = MPSC_DRIVER_NAME,
  1618. .dev_name = MPSC_DEV_NAME,
  1619. .major = MPSC_MAJOR,
  1620. .minor = MPSC_MINOR_START,
  1621. .nr = MPSC_NUM_CTLRS,
  1622. .cons = MPSC_CONSOLE,
  1623. };
  1624. static int mpsc_drv_map_regs(struct mpsc_port_info *pi,
  1625. struct platform_device *pd)
  1626. {
  1627. struct resource *r;
  1628. if ((r = platform_get_resource(pd, IORESOURCE_MEM, MPSC_BASE_ORDER))
  1629. && request_mem_region(r->start, MPSC_REG_BLOCK_SIZE,
  1630. "mpsc_regs")) {
  1631. pi->mpsc_base = ioremap(r->start, MPSC_REG_BLOCK_SIZE);
  1632. pi->mpsc_base_p = r->start;
  1633. } else {
  1634. mpsc_resource_err("MPSC base");
  1635. goto err;
  1636. }
  1637. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1638. MPSC_SDMA_BASE_ORDER))
  1639. && request_mem_region(r->start,
  1640. MPSC_SDMA_REG_BLOCK_SIZE, "sdma_regs")) {
  1641. pi->sdma_base = ioremap(r->start,MPSC_SDMA_REG_BLOCK_SIZE);
  1642. pi->sdma_base_p = r->start;
  1643. } else {
  1644. mpsc_resource_err("SDMA base");
  1645. goto err;
  1646. }
  1647. if ((r = platform_get_resource(pd,IORESOURCE_MEM,MPSC_BRG_BASE_ORDER))
  1648. && request_mem_region(r->start,
  1649. MPSC_BRG_REG_BLOCK_SIZE, "brg_regs")) {
  1650. pi->brg_base = ioremap(r->start, MPSC_BRG_REG_BLOCK_SIZE);
  1651. pi->brg_base_p = r->start;
  1652. } else {
  1653. mpsc_resource_err("BRG base");
  1654. goto err;
  1655. }
  1656. return 0;
  1657. err:
  1658. if (pi->sdma_base) {
  1659. iounmap(pi->sdma_base);
  1660. pi->sdma_base = NULL;
  1661. }
  1662. if (pi->mpsc_base) {
  1663. iounmap(pi->mpsc_base);
  1664. pi->mpsc_base = NULL;
  1665. }
  1666. return -ENOMEM;
  1667. }
  1668. static void mpsc_drv_unmap_regs(struct mpsc_port_info *pi)
  1669. {
  1670. if (pi->mpsc_base) {
  1671. iounmap(pi->mpsc_base);
  1672. release_mem_region(pi->mpsc_base_p, MPSC_REG_BLOCK_SIZE);
  1673. }
  1674. if (pi->sdma_base) {
  1675. iounmap(pi->sdma_base);
  1676. release_mem_region(pi->sdma_base_p, MPSC_SDMA_REG_BLOCK_SIZE);
  1677. }
  1678. if (pi->brg_base) {
  1679. iounmap(pi->brg_base);
  1680. release_mem_region(pi->brg_base_p, MPSC_BRG_REG_BLOCK_SIZE);
  1681. }
  1682. pi->mpsc_base = NULL;
  1683. pi->sdma_base = NULL;
  1684. pi->brg_base = NULL;
  1685. pi->mpsc_base_p = 0;
  1686. pi->sdma_base_p = 0;
  1687. pi->brg_base_p = 0;
  1688. }
  1689. static void mpsc_drv_get_platform_data(struct mpsc_port_info *pi,
  1690. struct platform_device *pd, int num)
  1691. {
  1692. struct mpsc_pdata *pdata;
  1693. pdata = dev_get_platdata(&pd->dev);
  1694. pi->port.uartclk = pdata->brg_clk_freq;
  1695. pi->port.iotype = UPIO_MEM;
  1696. pi->port.line = num;
  1697. pi->port.type = PORT_MPSC;
  1698. pi->port.fifosize = MPSC_TXBE_SIZE;
  1699. pi->port.membase = pi->mpsc_base;
  1700. pi->port.mapbase = (ulong)pi->mpsc_base;
  1701. pi->port.ops = &mpsc_pops;
  1702. pi->mirror_regs = pdata->mirror_regs;
  1703. pi->cache_mgmt = pdata->cache_mgmt;
  1704. pi->brg_can_tune = pdata->brg_can_tune;
  1705. pi->brg_clk_src = pdata->brg_clk_src;
  1706. pi->mpsc_max_idle = pdata->max_idle;
  1707. pi->default_baud = pdata->default_baud;
  1708. pi->default_bits = pdata->default_bits;
  1709. pi->default_parity = pdata->default_parity;
  1710. pi->default_flow = pdata->default_flow;
  1711. /* Initial values of mirrored regs */
  1712. pi->MPSC_CHR_1_m = pdata->chr_1_val;
  1713. pi->MPSC_CHR_2_m = pdata->chr_2_val;
  1714. pi->MPSC_CHR_10_m = pdata->chr_10_val;
  1715. pi->MPSC_MPCR_m = pdata->mpcr_val;
  1716. pi->BRG_BCR_m = pdata->bcr_val;
  1717. pi->shared_regs = &mpsc_shared_regs;
  1718. pi->port.irq = platform_get_irq(pd, 0);
  1719. }
  1720. static int mpsc_drv_probe(struct platform_device *dev)
  1721. {
  1722. struct mpsc_port_info *pi;
  1723. int rc;
  1724. dev_dbg(&dev->dev, "mpsc_drv_probe: Adding MPSC %d\n", dev->id);
  1725. if (dev->id >= MPSC_NUM_CTLRS)
  1726. return -ENODEV;
  1727. pi = &mpsc_ports[dev->id];
  1728. rc = mpsc_drv_map_regs(pi, dev);
  1729. if (rc)
  1730. return rc;
  1731. mpsc_drv_get_platform_data(pi, dev, dev->id);
  1732. pi->port.dev = &dev->dev;
  1733. rc = mpsc_make_ready(pi);
  1734. if (rc)
  1735. goto err_unmap;
  1736. spin_lock_init(&pi->tx_lock);
  1737. rc = uart_add_one_port(&mpsc_reg, &pi->port);
  1738. if (rc)
  1739. goto err_relport;
  1740. return 0;
  1741. err_relport:
  1742. mpsc_release_port(&pi->port);
  1743. err_unmap:
  1744. mpsc_drv_unmap_regs(pi);
  1745. return rc;
  1746. }
  1747. static struct platform_driver mpsc_driver = {
  1748. .probe = mpsc_drv_probe,
  1749. .driver = {
  1750. .name = MPSC_CTLR_NAME,
  1751. .suppress_bind_attrs = true,
  1752. },
  1753. };
  1754. static int __init mpsc_drv_init(void)
  1755. {
  1756. int rc;
  1757. printk(KERN_INFO "Serial: MPSC driver\n");
  1758. memset(mpsc_ports, 0, sizeof(mpsc_ports));
  1759. memset(&mpsc_shared_regs, 0, sizeof(mpsc_shared_regs));
  1760. rc = uart_register_driver(&mpsc_reg);
  1761. if (rc)
  1762. return rc;
  1763. rc = platform_driver_register(&mpsc_shared_driver);
  1764. if (rc)
  1765. goto err_unreg_uart;
  1766. rc = platform_driver_register(&mpsc_driver);
  1767. if (rc)
  1768. goto err_unreg_plat;
  1769. return 0;
  1770. err_unreg_plat:
  1771. platform_driver_unregister(&mpsc_shared_driver);
  1772. err_unreg_uart:
  1773. uart_unregister_driver(&mpsc_reg);
  1774. return rc;
  1775. }
  1776. device_initcall(mpsc_drv_init);
  1777. /*
  1778. MODULE_AUTHOR("Mark A. Greer <mgreer@mvista.com>");
  1779. MODULE_DESCRIPTION("Generic Marvell MPSC serial/UART driver");
  1780. MODULE_LICENSE("GPL");
  1781. */