|
@@ -239,6 +239,26 @@ static struct page *get_a_page(struct receive_queue *rq, gfp_t gfp_mask)
|
|
|
return p;
|
|
|
}
|
|
|
|
|
|
+static void virtqueue_napi_schedule(struct napi_struct *napi,
|
|
|
+ struct virtqueue *vq)
|
|
|
+{
|
|
|
+ if (napi_schedule_prep(napi)) {
|
|
|
+ virtqueue_disable_cb(vq);
|
|
|
+ __napi_schedule(napi);
|
|
|
+ }
|
|
|
+}
|
|
|
+
|
|
|
+static void virtqueue_napi_complete(struct napi_struct *napi,
|
|
|
+ struct virtqueue *vq, int processed)
|
|
|
+{
|
|
|
+ int opaque;
|
|
|
+
|
|
|
+ opaque = virtqueue_enable_cb_prepare(vq);
|
|
|
+ if (napi_complete_done(napi, processed) &&
|
|
|
+ unlikely(virtqueue_poll(vq, opaque)))
|
|
|
+ virtqueue_napi_schedule(napi, vq);
|
|
|
+}
|
|
|
+
|
|
|
static void skb_xmit_done(struct virtqueue *vq)
|
|
|
{
|
|
|
struct virtnet_info *vi = vq->vdev->priv;
|
|
@@ -936,27 +956,20 @@ static void skb_recv_done(struct virtqueue *rvq)
|
|
|
struct virtnet_info *vi = rvq->vdev->priv;
|
|
|
struct receive_queue *rq = &vi->rq[vq2rxq(rvq)];
|
|
|
|
|
|
- /* Schedule NAPI, Suppress further interrupts if successful. */
|
|
|
- if (napi_schedule_prep(&rq->napi)) {
|
|
|
- virtqueue_disable_cb(rvq);
|
|
|
- __napi_schedule(&rq->napi);
|
|
|
- }
|
|
|
+ virtqueue_napi_schedule(&rq->napi, rvq);
|
|
|
}
|
|
|
|
|
|
-static void virtnet_napi_enable(struct receive_queue *rq)
|
|
|
+static void virtnet_napi_enable(struct virtqueue *vq, struct napi_struct *napi)
|
|
|
{
|
|
|
- napi_enable(&rq->napi);
|
|
|
+ napi_enable(napi);
|
|
|
|
|
|
/* If all buffers were filled by other side before we napi_enabled, we
|
|
|
- * won't get another interrupt, so process any outstanding packets
|
|
|
- * now. virtnet_poll wants re-enable the queue, so we disable here.
|
|
|
- * We synchronize against interrupts via NAPI_STATE_SCHED */
|
|
|
- if (napi_schedule_prep(&rq->napi)) {
|
|
|
- virtqueue_disable_cb(rq->vq);
|
|
|
- local_bh_disable();
|
|
|
- __napi_schedule(&rq->napi);
|
|
|
- local_bh_enable();
|
|
|
- }
|
|
|
+ * won't get another interrupt, so process any outstanding packets now.
|
|
|
+ * Call local_bh_enable after to trigger softIRQ processing.
|
|
|
+ */
|
|
|
+ local_bh_disable();
|
|
|
+ virtqueue_napi_schedule(napi, vq);
|
|
|
+ local_bh_enable();
|
|
|
}
|
|
|
|
|
|
static void refill_work(struct work_struct *work)
|
|
@@ -971,7 +984,7 @@ static void refill_work(struct work_struct *work)
|
|
|
|
|
|
napi_disable(&rq->napi);
|
|
|
still_empty = !try_fill_recv(vi, rq, GFP_KERNEL);
|
|
|
- virtnet_napi_enable(rq);
|
|
|
+ virtnet_napi_enable(rq->vq, &rq->napi);
|
|
|
|
|
|
/* In theory, this can happen: if we don't get any buffers in
|
|
|
* we will *never* try to fill again.
|
|
@@ -1011,21 +1024,13 @@ static int virtnet_poll(struct napi_struct *napi, int budget)
|
|
|
{
|
|
|
struct receive_queue *rq =
|
|
|
container_of(napi, struct receive_queue, napi);
|
|
|
- unsigned int r, received;
|
|
|
+ unsigned int received;
|
|
|
|
|
|
received = virtnet_receive(rq, budget);
|
|
|
|
|
|
/* Out of packets? */
|
|
|
- if (received < budget) {
|
|
|
- r = virtqueue_enable_cb_prepare(rq->vq);
|
|
|
- if (napi_complete_done(napi, received)) {
|
|
|
- if (unlikely(virtqueue_poll(rq->vq, r)) &&
|
|
|
- napi_schedule_prep(napi)) {
|
|
|
- virtqueue_disable_cb(rq->vq);
|
|
|
- __napi_schedule(napi);
|
|
|
- }
|
|
|
- }
|
|
|
- }
|
|
|
+ if (received < budget)
|
|
|
+ virtqueue_napi_complete(napi, rq->vq, received);
|
|
|
|
|
|
return received;
|
|
|
}
|
|
@@ -1040,7 +1045,7 @@ static int virtnet_open(struct net_device *dev)
|
|
|
/* Make sure we have some buffers: if oom use wq. */
|
|
|
if (!try_fill_recv(vi, &vi->rq[i], GFP_KERNEL))
|
|
|
schedule_delayed_work(&vi->refill, 0);
|
|
|
- virtnet_napi_enable(&vi->rq[i]);
|
|
|
+ virtnet_napi_enable(vi->rq[i].vq, &vi->rq[i].napi);
|
|
|
}
|
|
|
|
|
|
return 0;
|
|
@@ -1747,7 +1752,7 @@ static int virtnet_restore_up(struct virtio_device *vdev)
|
|
|
schedule_delayed_work(&vi->refill, 0);
|
|
|
|
|
|
for (i = 0; i < vi->max_queue_pairs; i++)
|
|
|
- virtnet_napi_enable(&vi->rq[i]);
|
|
|
+ virtnet_napi_enable(vi->rq[i].vq, &vi->rq[i].napi);
|
|
|
}
|
|
|
|
|
|
netif_device_attach(vi->dev);
|