123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119 |
- // SPDX-License-Identifier: GPL-2.0
- /* XDP user-space ring structure
- * Copyright(c) 2018 Intel Corporation.
- */
- #include <linux/log2.h>
- #include <linux/slab.h>
- #include <linux/overflow.h>
- #include "xsk_queue.h"
- void xskq_set_umem(struct xsk_queue *q, u64 size, u64 chunk_mask)
- {
- if (!q)
- return;
- q->size = size;
- q->chunk_mask = chunk_mask;
- }
- static u32 xskq_umem_get_ring_size(struct xsk_queue *q)
- {
- return sizeof(struct xdp_umem_ring) + q->nentries * sizeof(u64);
- }
- static u32 xskq_rxtx_get_ring_size(struct xsk_queue *q)
- {
- return sizeof(struct xdp_ring) + q->nentries * sizeof(struct xdp_desc);
- }
- struct xsk_queue *xskq_create(u32 nentries, bool umem_queue)
- {
- struct xsk_queue *q;
- gfp_t gfp_flags;
- size_t size;
- q = kzalloc(sizeof(*q), GFP_KERNEL);
- if (!q)
- return NULL;
- q->nentries = nentries;
- q->ring_mask = nentries - 1;
- gfp_flags = GFP_KERNEL | __GFP_ZERO | __GFP_NOWARN |
- __GFP_COMP | __GFP_NORETRY;
- size = umem_queue ? xskq_umem_get_ring_size(q) :
- xskq_rxtx_get_ring_size(q);
- q->ring = (struct xdp_ring *)__get_free_pages(gfp_flags,
- get_order(size));
- if (!q->ring) {
- kfree(q);
- return NULL;
- }
- return q;
- }
- void xskq_destroy(struct xsk_queue *q)
- {
- if (!q)
- return;
- page_frag_free(q->ring);
- kfree(q);
- }
- struct xdp_umem_fq_reuse *xsk_reuseq_prepare(u32 nentries)
- {
- struct xdp_umem_fq_reuse *newq;
- /* Check for overflow */
- if (nentries > (u32)roundup_pow_of_two(nentries))
- return NULL;
- nentries = roundup_pow_of_two(nentries);
- newq = kvmalloc(struct_size(newq, handles, nentries), GFP_KERNEL);
- if (!newq)
- return NULL;
- memset(newq, 0, offsetof(typeof(*newq), handles));
- newq->nentries = nentries;
- return newq;
- }
- EXPORT_SYMBOL_GPL(xsk_reuseq_prepare);
- struct xdp_umem_fq_reuse *xsk_reuseq_swap(struct xdp_umem *umem,
- struct xdp_umem_fq_reuse *newq)
- {
- struct xdp_umem_fq_reuse *oldq = umem->fq_reuse;
- if (!oldq) {
- umem->fq_reuse = newq;
- return NULL;
- }
- if (newq->nentries < oldq->length)
- return newq;
- memcpy(newq->handles, oldq->handles,
- array_size(oldq->length, sizeof(u64)));
- newq->length = oldq->length;
- umem->fq_reuse = newq;
- return oldq;
- }
- EXPORT_SYMBOL_GPL(xsk_reuseq_swap);
- void xsk_reuseq_free(struct xdp_umem_fq_reuse *rq)
- {
- kvfree(rq);
- }
- EXPORT_SYMBOL_GPL(xsk_reuseq_free);
- void xsk_reuseq_destroy(struct xdp_umem *umem)
- {
- xsk_reuseq_free(umem->fq_reuse);
- umem->fq_reuse = NULL;
- }
|