Commit 9f78bf33 authored by Xuan Zhuo's avatar Xuan Zhuo Committed by David S. Miller

xsk: support use vaddr as ring

When we try to start AF_XDP on some machines with long running time, due
to the machine's memory fragmentation problem, there is no sufficient
contiguous physical memory that will cause the start failure.

If the size of the queue is 8 * 1024, then the size of the desc[] is
8 * 1024 * 8 = 16 * PAGE, but we also add struct xdp_ring size, so it is
16page+. This is necessary to apply for a 4-order memory. If there are a
lot of queues, it is difficult to these machine with long running time.

Here, that we actually waste 15 pages. 4-Order memory is 32 pages, but
we only use 17 pages.

This patch replaces __get_free_pages() by vmalloc() to allocate memory
to solve these problems.
Signed-off-by: default avatarXuan Zhuo <xuanzhuo@linux.alibaba.com>
Acked-by: default avatarMagnus Karlsson <magnus.karlsson@intel.com>
Reviewed-by: default avatarAlexander Lobakin <aleksander.lobakin@intel.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent b148d400
...@@ -1294,8 +1294,6 @@ static int xsk_mmap(struct file *file, struct socket *sock, ...@@ -1294,8 +1294,6 @@ static int xsk_mmap(struct file *file, struct socket *sock,
unsigned long size = vma->vm_end - vma->vm_start; unsigned long size = vma->vm_end - vma->vm_start;
struct xdp_sock *xs = xdp_sk(sock->sk); struct xdp_sock *xs = xdp_sk(sock->sk);
struct xsk_queue *q = NULL; struct xsk_queue *q = NULL;
unsigned long pfn;
struct page *qpg;
if (READ_ONCE(xs->state) != XSK_READY) if (READ_ONCE(xs->state) != XSK_READY)
return -EBUSY; return -EBUSY;
...@@ -1318,13 +1316,10 @@ static int xsk_mmap(struct file *file, struct socket *sock, ...@@ -1318,13 +1316,10 @@ static int xsk_mmap(struct file *file, struct socket *sock,
/* Matches the smp_wmb() in xsk_init_queue */ /* Matches the smp_wmb() in xsk_init_queue */
smp_rmb(); smp_rmb();
qpg = virt_to_head_page(q->ring); if (size > q->ring_vmalloc_size)
if (size > page_size(qpg))
return -EINVAL; return -EINVAL;
pfn = virt_to_phys(q->ring) >> PAGE_SHIFT; return remap_vmalloc_range(vma, q->ring, 0);
return remap_pfn_range(vma, vma->vm_start, pfn,
size, vma->vm_page_prot);
} }
static int xsk_notifier(struct notifier_block *this, static int xsk_notifier(struct notifier_block *this,
......
...@@ -6,6 +6,7 @@ ...@@ -6,6 +6,7 @@
#include <linux/log2.h> #include <linux/log2.h>
#include <linux/slab.h> #include <linux/slab.h>
#include <linux/overflow.h> #include <linux/overflow.h>
#include <linux/vmalloc.h>
#include <net/xdp_sock_drv.h> #include <net/xdp_sock_drv.h>
#include "xsk_queue.h" #include "xsk_queue.h"
...@@ -23,7 +24,6 @@ static size_t xskq_get_ring_size(struct xsk_queue *q, bool umem_queue) ...@@ -23,7 +24,6 @@ static size_t xskq_get_ring_size(struct xsk_queue *q, bool umem_queue)
struct xsk_queue *xskq_create(u32 nentries, bool umem_queue) struct xsk_queue *xskq_create(u32 nentries, bool umem_queue)
{ {
struct xsk_queue *q; struct xsk_queue *q;
gfp_t gfp_flags;
size_t size; size_t size;
q = kzalloc(sizeof(*q), GFP_KERNEL); q = kzalloc(sizeof(*q), GFP_KERNEL);
...@@ -33,17 +33,16 @@ struct xsk_queue *xskq_create(u32 nentries, bool umem_queue) ...@@ -33,17 +33,16 @@ struct xsk_queue *xskq_create(u32 nentries, bool umem_queue)
q->nentries = nentries; q->nentries = nentries;
q->ring_mask = nentries - 1; q->ring_mask = nentries - 1;
gfp_flags = GFP_KERNEL | __GFP_ZERO | __GFP_NOWARN |
__GFP_COMP | __GFP_NORETRY;
size = xskq_get_ring_size(q, umem_queue); size = xskq_get_ring_size(q, umem_queue);
size = PAGE_ALIGN(size);
q->ring = (struct xdp_ring *)__get_free_pages(gfp_flags, q->ring = vmalloc_user(size);
get_order(size));
if (!q->ring) { if (!q->ring) {
kfree(q); kfree(q);
return NULL; return NULL;
} }
q->ring_vmalloc_size = size;
return q; return q;
} }
...@@ -52,6 +51,6 @@ void xskq_destroy(struct xsk_queue *q) ...@@ -52,6 +51,6 @@ void xskq_destroy(struct xsk_queue *q)
if (!q) if (!q)
return; return;
page_frag_free(q->ring); vfree(q->ring);
kfree(q); kfree(q);
} }
...@@ -45,6 +45,7 @@ struct xsk_queue { ...@@ -45,6 +45,7 @@ struct xsk_queue {
struct xdp_ring *ring; struct xdp_ring *ring;
u64 invalid_descs; u64 invalid_descs;
u64 queue_empty_descs; u64 queue_empty_descs;
size_t ring_vmalloc_size;
}; };
/* The structure of the shared state of the rings are a simple /* The structure of the shared state of the rings are a simple
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment