// SPDX-License-Identifier: GPL-2.0 /* XDP user-space ring structure * Copyright(c) 2018 Intel Corporation. */ #include #include #include #include "xsk_queue.h" void xskq_set_umem(struct xsk_queue *q, u64 size, u64 chunk_mask) { if (!q) return; q->size = size; q->chunk_mask = chunk_mask; } static u32 xskq_umem_get_ring_size(struct xsk_queue *q) { return sizeof(struct xdp_umem_ring) + q->nentries * sizeof(u64); } static u32 xskq_rxtx_get_ring_size(struct xsk_queue *q) { return sizeof(struct xdp_ring) + q->nentries * sizeof(struct xdp_desc); } struct xsk_queue *xskq_create(u32 nentries, bool umem_queue) { struct xsk_queue *q; gfp_t gfp_flags; size_t size; q = kzalloc(sizeof(*q), GFP_KERNEL); if (!q) return NULL; q->nentries = nentries; q->ring_mask = nentries - 1; gfp_flags = GFP_KERNEL | __GFP_ZERO | __GFP_NOWARN | __GFP_COMP | __GFP_NORETRY; size = umem_queue ? xskq_umem_get_ring_size(q) : xskq_rxtx_get_ring_size(q); q->ring = (struct xdp_ring *)__get_free_pages(gfp_flags, get_order(size)); if (!q->ring) { kfree(q); return NULL; } return q; } void xskq_destroy(struct xsk_queue *q) { if (!q) return; page_frag_free(q->ring); kfree(q); } struct xdp_umem_fq_reuse *xsk_reuseq_prepare(u32 nentries) { struct xdp_umem_fq_reuse *newq; /* Check for overflow */ if (nentries > (u32)roundup_pow_of_two(nentries)) return NULL; nentries = roundup_pow_of_two(nentries); newq = kvmalloc(struct_size(newq, handles, nentries), GFP_KERNEL); if (!newq) return NULL; memset(newq, 0, offsetof(typeof(*newq), handles)); newq->nentries = nentries; return newq; } EXPORT_SYMBOL_GPL(xsk_reuseq_prepare); struct xdp_umem_fq_reuse *xsk_reuseq_swap(struct xdp_umem *umem, struct xdp_umem_fq_reuse *newq) { struct xdp_umem_fq_reuse *oldq = umem->fq_reuse; if (!oldq) { umem->fq_reuse = newq; return NULL; } if (newq->nentries < oldq->length) return newq; memcpy(newq->handles, oldq->handles, array_size(oldq->length, sizeof(u64))); newq->length = oldq->length; umem->fq_reuse = newq; return oldq; } EXPORT_SYMBOL_GPL(xsk_reuseq_swap); void xsk_reuseq_free(struct xdp_umem_fq_reuse *rq) { kvfree(rq); } EXPORT_SYMBOL_GPL(xsk_reuseq_free); void xsk_reuseq_destroy(struct xdp_umem *umem) { xsk_reuseq_free(umem->fq_reuse); umem->fq_reuse = NULL; }