|
|
@@ -9,7 +9,7 @@
|
|
|
* Maintainer: Kumar Gala
|
|
|
* Modifier: Sandeep Gopalpet <sandeep.kumar@freescale.com>
|
|
|
*
|
|
|
- * Copyright 2002-2009, 2011 Freescale Semiconductor, Inc.
|
|
|
+ * Copyright 2002-2009, 2011-2013 Freescale Semiconductor, Inc.
|
|
|
* Copyright 2007 MontaVista Software, Inc.
|
|
|
*
|
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
|
@@ -511,7 +511,43 @@ void unlock_tx_qs(struct gfar_private *priv)
|
|
|
spin_unlock(&priv->tx_queue[i]->txlock);
|
|
|
}
|
|
|
|
|
|
-static void free_tx_pointers(struct gfar_private *priv)
|
|
|
+static int gfar_alloc_tx_queues(struct gfar_private *priv)
|
|
|
+{
|
|
|
+ int i;
|
|
|
+
|
|
|
+ for (i = 0; i < priv->num_tx_queues; i++) {
|
|
|
+ priv->tx_queue[i] = kzalloc(sizeof(struct gfar_priv_tx_q),
|
|
|
+ GFP_KERNEL);
|
|
|
+ if (!priv->tx_queue[i])
|
|
|
+ return -ENOMEM;
|
|
|
+
|
|
|
+ priv->tx_queue[i]->tx_skbuff = NULL;
|
|
|
+ priv->tx_queue[i]->qindex = i;
|
|
|
+ priv->tx_queue[i]->dev = priv->ndev;
|
|
|
+ spin_lock_init(&(priv->tx_queue[i]->txlock));
|
|
|
+ }
|
|
|
+ return 0;
|
|
|
+}
|
|
|
+
|
|
|
+static int gfar_alloc_rx_queues(struct gfar_private *priv)
|
|
|
+{
|
|
|
+ int i;
|
|
|
+
|
|
|
+ for (i = 0; i < priv->num_rx_queues; i++) {
|
|
|
+ priv->rx_queue[i] = kzalloc(sizeof(struct gfar_priv_rx_q),
|
|
|
+ GFP_KERNEL);
|
|
|
+ if (!priv->rx_queue[i])
|
|
|
+ return -ENOMEM;
|
|
|
+
|
|
|
+ priv->rx_queue[i]->rx_skbuff = NULL;
|
|
|
+ priv->rx_queue[i]->qindex = i;
|
|
|
+ priv->rx_queue[i]->dev = priv->ndev;
|
|
|
+ spin_lock_init(&(priv->rx_queue[i]->rxlock));
|
|
|
+ }
|
|
|
+ return 0;
|
|
|
+}
|
|
|
+
|
|
|
+static void gfar_free_tx_queues(struct gfar_private *priv)
|
|
|
{
|
|
|
int i;
|
|
|
|
|
|
@@ -519,7 +555,7 @@ static void free_tx_pointers(struct gfar_private *priv)
|
|
|
kfree(priv->tx_queue[i]);
|
|
|
}
|
|
|
|
|
|
-static void free_rx_pointers(struct gfar_private *priv)
|
|
|
+static void gfar_free_rx_queues(struct gfar_private *priv)
|
|
|
{
|
|
|
int i;
|
|
|
|
|
|
@@ -608,6 +644,30 @@ static int gfar_parse_group(struct device_node *np,
|
|
|
grp->rx_bit_map = 0xFF;
|
|
|
grp->tx_bit_map = 0xFF;
|
|
|
}
|
|
|
+
|
|
|
+ /* bit_map's MSB is q0 (from q0 to q7) but, for_each_set_bit parses
|
|
|
+ * right to left, so we need to revert the 8 bits to get the q index
|
|
|
+ */
|
|
|
+ grp->rx_bit_map = bitrev8(grp->rx_bit_map);
|
|
|
+ grp->tx_bit_map = bitrev8(grp->tx_bit_map);
|
|
|
+
|
|
|
+ /* Calculate RSTAT, TSTAT, RQUEUE and TQUEUE values,
|
|
|
+ * also assign queues to groups
|
|
|
+ */
|
|
|
+ for_each_set_bit(i, &grp->rx_bit_map, priv->num_rx_queues) {
|
|
|
+ grp->num_rx_queues++;
|
|
|
+ grp->rstat |= (RSTAT_CLEAR_RHALT >> i);
|
|
|
+ priv->rqueue |= ((RQUEUE_EN0 | RQUEUE_EX0) >> i);
|
|
|
+ priv->rx_queue[i]->grp = grp;
|
|
|
+ }
|
|
|
+
|
|
|
+ for_each_set_bit(i, &grp->tx_bit_map, priv->num_tx_queues) {
|
|
|
+ grp->num_tx_queues++;
|
|
|
+ grp->tstat |= (TSTAT_CLEAR_THALT >> i);
|
|
|
+ priv->tqueue |= (TQUEUE_EN0 >> i);
|
|
|
+ priv->tx_queue[i]->grp = grp;
|
|
|
+ }
|
|
|
+
|
|
|
priv->num_grps++;
|
|
|
|
|
|
return 0;
|
|
|
@@ -664,7 +724,14 @@ static int gfar_of_init(struct platform_device *ofdev, struct net_device **pdev)
|
|
|
priv->num_tx_queues = num_tx_qs;
|
|
|
netif_set_real_num_rx_queues(dev, num_rx_qs);
|
|
|
priv->num_rx_queues = num_rx_qs;
|
|
|
- priv->num_grps = 0x0;
|
|
|
+
|
|
|
+ err = gfar_alloc_tx_queues(priv);
|
|
|
+ if (err)
|
|
|
+ goto tx_alloc_failed;
|
|
|
+
|
|
|
+ err = gfar_alloc_rx_queues(priv);
|
|
|
+ if (err)
|
|
|
+ goto rx_alloc_failed;
|
|
|
|
|
|
/* Init Rx queue filer rule set linked list */
|
|
|
INIT_LIST_HEAD(&priv->rx_list.list);
|
|
|
@@ -691,38 +758,6 @@ static int gfar_of_init(struct platform_device *ofdev, struct net_device **pdev)
|
|
|
goto err_grp_init;
|
|
|
}
|
|
|
|
|
|
- for (i = 0; i < priv->num_tx_queues; i++)
|
|
|
- priv->tx_queue[i] = NULL;
|
|
|
- for (i = 0; i < priv->num_rx_queues; i++)
|
|
|
- priv->rx_queue[i] = NULL;
|
|
|
-
|
|
|
- for (i = 0; i < priv->num_tx_queues; i++) {
|
|
|
- priv->tx_queue[i] = kzalloc(sizeof(struct gfar_priv_tx_q),
|
|
|
- GFP_KERNEL);
|
|
|
- if (!priv->tx_queue[i]) {
|
|
|
- err = -ENOMEM;
|
|
|
- goto tx_alloc_failed;
|
|
|
- }
|
|
|
- priv->tx_queue[i]->tx_skbuff = NULL;
|
|
|
- priv->tx_queue[i]->qindex = i;
|
|
|
- priv->tx_queue[i]->dev = dev;
|
|
|
- spin_lock_init(&(priv->tx_queue[i]->txlock));
|
|
|
- }
|
|
|
-
|
|
|
- for (i = 0; i < priv->num_rx_queues; i++) {
|
|
|
- priv->rx_queue[i] = kzalloc(sizeof(struct gfar_priv_rx_q),
|
|
|
- GFP_KERNEL);
|
|
|
- if (!priv->rx_queue[i]) {
|
|
|
- err = -ENOMEM;
|
|
|
- goto rx_alloc_failed;
|
|
|
- }
|
|
|
- priv->rx_queue[i]->rx_skbuff = NULL;
|
|
|
- priv->rx_queue[i]->qindex = i;
|
|
|
- priv->rx_queue[i]->dev = dev;
|
|
|
- spin_lock_init(&(priv->rx_queue[i]->rxlock));
|
|
|
- }
|
|
|
-
|
|
|
-
|
|
|
stash = of_get_property(np, "bd-stash", NULL);
|
|
|
|
|
|
if (stash) {
|
|
|
@@ -784,12 +819,12 @@ static int gfar_of_init(struct platform_device *ofdev, struct net_device **pdev)
|
|
|
|
|
|
return 0;
|
|
|
|
|
|
-rx_alloc_failed:
|
|
|
- free_rx_pointers(priv);
|
|
|
-tx_alloc_failed:
|
|
|
- free_tx_pointers(priv);
|
|
|
err_grp_init:
|
|
|
unmap_group_regs(priv);
|
|
|
+rx_alloc_failed:
|
|
|
+ gfar_free_rx_queues(priv);
|
|
|
+tx_alloc_failed:
|
|
|
+ gfar_free_tx_queues(priv);
|
|
|
free_gfar_dev(priv);
|
|
|
return err;
|
|
|
}
|
|
|
@@ -875,19 +910,6 @@ static int gfar_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
|
|
|
return phy_mii_ioctl(priv->phydev, rq, cmd);
|
|
|
}
|
|
|
|
|
|
-static unsigned int reverse_bitmap(unsigned int bit_map, unsigned int max_qs)
|
|
|
-{
|
|
|
- unsigned int new_bit_map = 0x0;
|
|
|
- int mask = 0x1 << (max_qs - 1), i;
|
|
|
-
|
|
|
- for (i = 0; i < max_qs; i++) {
|
|
|
- if (bit_map & mask)
|
|
|
- new_bit_map = new_bit_map + (1 << i);
|
|
|
- mask = mask >> 0x1;
|
|
|
- }
|
|
|
- return new_bit_map;
|
|
|
-}
|
|
|
-
|
|
|
static u32 cluster_entry_per_class(struct gfar_private *priv, u32 rqfar,
|
|
|
u32 class)
|
|
|
{
|
|
|
@@ -1005,19 +1027,88 @@ static void gfar_detect_errata(struct gfar_private *priv)
|
|
|
priv->errata);
|
|
|
}
|
|
|
|
|
|
+static void gfar_hw_init(struct gfar_private *priv)
|
|
|
+{
|
|
|
+ struct gfar __iomem *regs = priv->gfargrp[0].regs;
|
|
|
+ u32 tempval;
|
|
|
+
|
|
|
+ /* Reset MAC layer */
|
|
|
+ gfar_write(®s->maccfg1, MACCFG1_SOFT_RESET);
|
|
|
+
|
|
|
+ /* We need to delay at least 3 TX clocks */
|
|
|
+ udelay(2);
|
|
|
+
|
|
|
+ /* the soft reset bit is not self-resetting, so we need to
|
|
|
+ * clear it before resuming normal operation
|
|
|
+ */
|
|
|
+ gfar_write(®s->maccfg1, 0);
|
|
|
+
|
|
|
+ /* Initialize MACCFG2. */
|
|
|
+ tempval = MACCFG2_INIT_SETTINGS;
|
|
|
+ if (gfar_has_errata(priv, GFAR_ERRATA_74))
|
|
|
+ tempval |= MACCFG2_HUGEFRAME | MACCFG2_LENGTHCHECK;
|
|
|
+ gfar_write(®s->maccfg2, tempval);
|
|
|
+
|
|
|
+ /* Initialize ECNTRL */
|
|
|
+ gfar_write(®s->ecntrl, ECNTRL_INIT_SETTINGS);
|
|
|
+
|
|
|
+ /* Program the interrupt steering regs, only for MG devices */
|
|
|
+ if (priv->num_grps > 1)
|
|
|
+ gfar_write_isrg(priv);
|
|
|
+
|
|
|
+ /* Enable all Rx/Tx queues after MAC reset */
|
|
|
+ gfar_write(®s->rqueue, priv->rqueue);
|
|
|
+ gfar_write(®s->tqueue, priv->tqueue);
|
|
|
+}
|
|
|
+
|
|
|
+static void __init gfar_init_addr_hash_table(struct gfar_private *priv)
|
|
|
+{
|
|
|
+ struct gfar __iomem *regs = priv->gfargrp[0].regs;
|
|
|
+
|
|
|
+ if (priv->device_flags & FSL_GIANFAR_DEV_HAS_EXTENDED_HASH) {
|
|
|
+ priv->extended_hash = 1;
|
|
|
+ priv->hash_width = 9;
|
|
|
+
|
|
|
+ priv->hash_regs[0] = ®s->igaddr0;
|
|
|
+ priv->hash_regs[1] = ®s->igaddr1;
|
|
|
+ priv->hash_regs[2] = ®s->igaddr2;
|
|
|
+ priv->hash_regs[3] = ®s->igaddr3;
|
|
|
+ priv->hash_regs[4] = ®s->igaddr4;
|
|
|
+ priv->hash_regs[5] = ®s->igaddr5;
|
|
|
+ priv->hash_regs[6] = ®s->igaddr6;
|
|
|
+ priv->hash_regs[7] = ®s->igaddr7;
|
|
|
+ priv->hash_regs[8] = ®s->gaddr0;
|
|
|
+ priv->hash_regs[9] = ®s->gaddr1;
|
|
|
+ priv->hash_regs[10] = ®s->gaddr2;
|
|
|
+ priv->hash_regs[11] = ®s->gaddr3;
|
|
|
+ priv->hash_regs[12] = ®s->gaddr4;
|
|
|
+ priv->hash_regs[13] = ®s->gaddr5;
|
|
|
+ priv->hash_regs[14] = ®s->gaddr6;
|
|
|
+ priv->hash_regs[15] = ®s->gaddr7;
|
|
|
+
|
|
|
+ } else {
|
|
|
+ priv->extended_hash = 0;
|
|
|
+ priv->hash_width = 8;
|
|
|
+
|
|
|
+ priv->hash_regs[0] = ®s->gaddr0;
|
|
|
+ priv->hash_regs[1] = ®s->gaddr1;
|
|
|
+ priv->hash_regs[2] = ®s->gaddr2;
|
|
|
+ priv->hash_regs[3] = ®s->gaddr3;
|
|
|
+ priv->hash_regs[4] = ®s->gaddr4;
|
|
|
+ priv->hash_regs[5] = ®s->gaddr5;
|
|
|
+ priv->hash_regs[6] = ®s->gaddr6;
|
|
|
+ priv->hash_regs[7] = ®s->gaddr7;
|
|
|
+ }
|
|
|
+}
|
|
|
+
|
|
|
/* Set up the ethernet device structure, private data,
|
|
|
* and anything else we need before we start
|
|
|
*/
|
|
|
static int gfar_probe(struct platform_device *ofdev)
|
|
|
{
|
|
|
- u32 tempval;
|
|
|
struct net_device *dev = NULL;
|
|
|
struct gfar_private *priv = NULL;
|
|
|
- struct gfar __iomem *regs = NULL;
|
|
|
- int err = 0, i, grp_idx = 0;
|
|
|
- u32 rstat = 0, tstat = 0, rqueue = 0, tqueue = 0;
|
|
|
- u32 isrg = 0;
|
|
|
- u32 __iomem *baddr;
|
|
|
+ int err = 0, i;
|
|
|
|
|
|
err = gfar_of_init(ofdev, &dev);
|
|
|
|
|
|
@@ -1034,7 +1125,6 @@ static int gfar_probe(struct platform_device *ofdev)
|
|
|
INIT_WORK(&priv->reset_task, gfar_reset_task);
|
|
|
|
|
|
platform_set_drvdata(ofdev, priv);
|
|
|
- regs = priv->gfargrp[0].regs;
|
|
|
|
|
|
gfar_detect_errata(priv);
|
|
|
|
|
|
@@ -1043,33 +1133,10 @@ static int gfar_probe(struct platform_device *ofdev)
|
|
|
*/
|
|
|
gfar_halt(dev);
|
|
|
|
|
|
- /* Reset MAC layer */
|
|
|
- gfar_write(®s->maccfg1, MACCFG1_SOFT_RESET);
|
|
|
-
|
|
|
- /* We need to delay at least 3 TX clocks */
|
|
|
- udelay(2);
|
|
|
-
|
|
|
- tempval = 0;
|
|
|
- if (!priv->pause_aneg_en && priv->tx_pause_en)
|
|
|
- tempval |= MACCFG1_TX_FLOW;
|
|
|
- if (!priv->pause_aneg_en && priv->rx_pause_en)
|
|
|
- tempval |= MACCFG1_RX_FLOW;
|
|
|
- /* the soft reset bit is not self-resetting, so we need to
|
|
|
- * clear it before resuming normal operation
|
|
|
- */
|
|
|
- gfar_write(®s->maccfg1, tempval);
|
|
|
-
|
|
|
- /* Initialize MACCFG2. */
|
|
|
- tempval = MACCFG2_INIT_SETTINGS;
|
|
|
- if (gfar_has_errata(priv, GFAR_ERRATA_74))
|
|
|
- tempval |= MACCFG2_HUGEFRAME | MACCFG2_LENGTHCHECK;
|
|
|
- gfar_write(®s->maccfg2, tempval);
|
|
|
-
|
|
|
- /* Initialize ECNTRL */
|
|
|
- gfar_write(®s->ecntrl, ECNTRL_INIT_SETTINGS);
|
|
|
+ gfar_hw_init(priv);
|
|
|
|
|
|
/* Set the dev->base_addr to the gfar reg region */
|
|
|
- dev->base_addr = (unsigned long) regs;
|
|
|
+ dev->base_addr = (unsigned long) priv->gfargrp[0].regs;
|
|
|
|
|
|
/* Fill in the dev structure */
|
|
|
dev->watchdog_timeo = TX_TIMEOUT;
|
|
|
@@ -1099,40 +1166,7 @@ static int gfar_probe(struct platform_device *ofdev)
|
|
|
dev->features |= NETIF_F_HW_VLAN_CTAG_RX;
|
|
|
}
|
|
|
|
|
|
- if (priv->device_flags & FSL_GIANFAR_DEV_HAS_EXTENDED_HASH) {
|
|
|
- priv->extended_hash = 1;
|
|
|
- priv->hash_width = 9;
|
|
|
-
|
|
|
- priv->hash_regs[0] = ®s->igaddr0;
|
|
|
- priv->hash_regs[1] = ®s->igaddr1;
|
|
|
- priv->hash_regs[2] = ®s->igaddr2;
|
|
|
- priv->hash_regs[3] = ®s->igaddr3;
|
|
|
- priv->hash_regs[4] = ®s->igaddr4;
|
|
|
- priv->hash_regs[5] = ®s->igaddr5;
|
|
|
- priv->hash_regs[6] = ®s->igaddr6;
|
|
|
- priv->hash_regs[7] = ®s->igaddr7;
|
|
|
- priv->hash_regs[8] = ®s->gaddr0;
|
|
|
- priv->hash_regs[9] = ®s->gaddr1;
|
|
|
- priv->hash_regs[10] = ®s->gaddr2;
|
|
|
- priv->hash_regs[11] = ®s->gaddr3;
|
|
|
- priv->hash_regs[12] = ®s->gaddr4;
|
|
|
- priv->hash_regs[13] = ®s->gaddr5;
|
|
|
- priv->hash_regs[14] = ®s->gaddr6;
|
|
|
- priv->hash_regs[15] = ®s->gaddr7;
|
|
|
-
|
|
|
- } else {
|
|
|
- priv->extended_hash = 0;
|
|
|
- priv->hash_width = 8;
|
|
|
-
|
|
|
- priv->hash_regs[0] = ®s->gaddr0;
|
|
|
- priv->hash_regs[1] = ®s->gaddr1;
|
|
|
- priv->hash_regs[2] = ®s->gaddr2;
|
|
|
- priv->hash_regs[3] = ®s->gaddr3;
|
|
|
- priv->hash_regs[4] = ®s->gaddr4;
|
|
|
- priv->hash_regs[5] = ®s->gaddr5;
|
|
|
- priv->hash_regs[6] = ®s->gaddr6;
|
|
|
- priv->hash_regs[7] = ®s->gaddr7;
|
|
|
- }
|
|
|
+ gfar_init_addr_hash_table(priv);
|
|
|
|
|
|
if (priv->device_flags & FSL_GIANFAR_DEV_HAS_PADDING)
|
|
|
priv->padding = DEFAULT_PADDING;
|
|
|
@@ -1143,59 +1177,6 @@ static int gfar_probe(struct platform_device *ofdev)
|
|
|
priv->device_flags & FSL_GIANFAR_DEV_HAS_TIMER)
|
|
|
dev->needed_headroom = GMAC_FCB_LEN;
|
|
|
|
|
|
- /* Program the isrg regs only if number of grps > 1 */
|
|
|
- if (priv->num_grps > 1) {
|
|
|
- baddr = ®s->isrg0;
|
|
|
- for (i = 0; i < priv->num_grps; i++) {
|
|
|
- isrg |= (priv->gfargrp[i].rx_bit_map << ISRG_SHIFT_RX);
|
|
|
- isrg |= (priv->gfargrp[i].tx_bit_map << ISRG_SHIFT_TX);
|
|
|
- gfar_write(baddr, isrg);
|
|
|
- baddr++;
|
|
|
- isrg = 0x0;
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
- /* Need to reverse the bit maps as bit_map's MSB is q0
|
|
|
- * but, for_each_set_bit parses from right to left, which
|
|
|
- * basically reverses the queue numbers
|
|
|
- */
|
|
|
- for (i = 0; i< priv->num_grps; i++) {
|
|
|
- priv->gfargrp[i].tx_bit_map =
|
|
|
- reverse_bitmap(priv->gfargrp[i].tx_bit_map, MAX_TX_QS);
|
|
|
- priv->gfargrp[i].rx_bit_map =
|
|
|
- reverse_bitmap(priv->gfargrp[i].rx_bit_map, MAX_RX_QS);
|
|
|
- }
|
|
|
-
|
|
|
- /* Calculate RSTAT, TSTAT, RQUEUE and TQUEUE values,
|
|
|
- * also assign queues to groups
|
|
|
- */
|
|
|
- for (grp_idx = 0; grp_idx < priv->num_grps; grp_idx++) {
|
|
|
- priv->gfargrp[grp_idx].num_rx_queues = 0x0;
|
|
|
-
|
|
|
- for_each_set_bit(i, &priv->gfargrp[grp_idx].rx_bit_map,
|
|
|
- priv->num_rx_queues) {
|
|
|
- priv->gfargrp[grp_idx].num_rx_queues++;
|
|
|
- priv->rx_queue[i]->grp = &priv->gfargrp[grp_idx];
|
|
|
- rstat = rstat | (RSTAT_CLEAR_RHALT >> i);
|
|
|
- rqueue = rqueue | ((RQUEUE_EN0 | RQUEUE_EX0) >> i);
|
|
|
- }
|
|
|
- priv->gfargrp[grp_idx].num_tx_queues = 0x0;
|
|
|
-
|
|
|
- for_each_set_bit(i, &priv->gfargrp[grp_idx].tx_bit_map,
|
|
|
- priv->num_tx_queues) {
|
|
|
- priv->gfargrp[grp_idx].num_tx_queues++;
|
|
|
- priv->tx_queue[i]->grp = &priv->gfargrp[grp_idx];
|
|
|
- tstat = tstat | (TSTAT_CLEAR_THALT >> i);
|
|
|
- tqueue = tqueue | (TQUEUE_EN0 >> i);
|
|
|
- }
|
|
|
- priv->gfargrp[grp_idx].rstat = rstat;
|
|
|
- priv->gfargrp[grp_idx].tstat = tstat;
|
|
|
- rstat = tstat =0;
|
|
|
- }
|
|
|
-
|
|
|
- gfar_write(®s->rqueue, rqueue);
|
|
|
- gfar_write(®s->tqueue, tqueue);
|
|
|
-
|
|
|
priv->rx_buffer_size = DEFAULT_RX_BUFFER_SIZE;
|
|
|
|
|
|
/* Initializing some of the rx/tx queue level parameters */
|
|
|
@@ -1272,8 +1253,8 @@ static int gfar_probe(struct platform_device *ofdev)
|
|
|
|
|
|
register_fail:
|
|
|
unmap_group_regs(priv);
|
|
|
- free_tx_pointers(priv);
|
|
|
- free_rx_pointers(priv);
|
|
|
+ gfar_free_rx_queues(priv);
|
|
|
+ gfar_free_tx_queues(priv);
|
|
|
if (priv->phy_node)
|
|
|
of_node_put(priv->phy_node);
|
|
|
if (priv->tbi_node)
|
|
|
@@ -1293,6 +1274,8 @@ static int gfar_remove(struct platform_device *ofdev)
|
|
|
|
|
|
unregister_netdev(priv->ndev);
|
|
|
unmap_group_regs(priv);
|
|
|
+ gfar_free_rx_queues(priv);
|
|
|
+ gfar_free_tx_queues(priv);
|
|
|
free_gfar_dev(priv);
|
|
|
|
|
|
return 0;
|