Commit 5511d781 authored by Mitko Haralanov's avatar Mitko Haralanov Committed by Doug Ledford

IB/hfi1: Add SDMA cache eviction algorithm

This commit adds a cache eviction algorithm for the SDMA
user buffer cache.

Besides the interval RB tree used for node lookup, the cache
nodes are also arranged in a doubly-linked list. When a node is
used, it is put at the beginning of the list. Less frequently
used nodes naturally move to the tail of the list.

When the cache limit is reached, the eviction code starts
traversing the linked list in reverse, freeing buffers until
enough space has been freed to fit the new user buffer. This
guarantees that only the least used cache nodes will be removed
from the cache.
Reviewed-by: default avatarDennis Dalessandro <dennis.dalessandro@intel.com>
Reviewed-by: default avatarDean Luick <dean.luick@intel.com>
Signed-off-by: default avatarMitko Haralanov <mitko.haralanov@intel.com>
Signed-off-by: default avatarJubin John <jubin.john@intel.com>
Signed-off-by: default avatarDoug Ledford <dledford@redhat.com>
parent a7922f7d
...@@ -183,6 +183,8 @@ struct user_sdma_iovec { ...@@ -183,6 +183,8 @@ struct user_sdma_iovec {
struct sdma_mmu_node { struct sdma_mmu_node {
struct mmu_rb_node rb; struct mmu_rb_node rb;
struct list_head list;
struct hfi1_user_sdma_pkt_q *pq;
atomic_t refcount; atomic_t refcount;
struct page **pages; struct page **pages;
unsigned npages; unsigned npages;
...@@ -397,6 +399,8 @@ int hfi1_user_sdma_alloc_queues(struct hfi1_ctxtdata *uctxt, struct file *fp) ...@@ -397,6 +399,8 @@ int hfi1_user_sdma_alloc_queues(struct hfi1_ctxtdata *uctxt, struct file *fp)
atomic_set(&pq->n_reqs, 0); atomic_set(&pq->n_reqs, 0);
init_waitqueue_head(&pq->wait); init_waitqueue_head(&pq->wait);
pq->sdma_rb_root = RB_ROOT; pq->sdma_rb_root = RB_ROOT;
INIT_LIST_HEAD(&pq->evict);
spin_lock_init(&pq->evict_lock);
iowait_init(&pq->busy, 0, NULL, defer_packet_queue, iowait_init(&pq->busy, 0, NULL, defer_packet_queue,
activate_packet_queue, NULL); activate_packet_queue, NULL);
...@@ -1027,9 +1031,33 @@ static inline int num_user_pages(const struct iovec *iov) ...@@ -1027,9 +1031,33 @@ static inline int num_user_pages(const struct iovec *iov)
return 1 + ((epage - spage) >> PAGE_SHIFT); return 1 + ((epage - spage) >> PAGE_SHIFT);
} }
/* Caller must hold pq->evict_lock */
static u32 sdma_cache_evict(struct hfi1_user_sdma_pkt_q *pq, u32 npages)
{
u32 cleared = 0;
struct sdma_mmu_node *node, *ptr;
list_for_each_entry_safe_reverse(node, ptr, &pq->evict, list) {
/* Make sure that no one is still using the node. */
if (!atomic_read(&node->refcount)) {
/*
* Need to use the page count now as the remove callback
* will free the node.
*/
cleared += node->npages;
spin_unlock(&pq->evict_lock);
hfi1_mmu_rb_remove(&pq->sdma_rb_root, &node->rb);
spin_lock(&pq->evict_lock);
if (cleared >= npages)
break;
}
}
return cleared;
}
static int pin_vector_pages(struct user_sdma_request *req, static int pin_vector_pages(struct user_sdma_request *req,
struct user_sdma_iovec *iovec) { struct user_sdma_iovec *iovec) {
int ret = 0, pinned, npages; int ret = 0, pinned, npages, cleared;
struct page **pages; struct page **pages;
struct hfi1_user_sdma_pkt_q *pq = req->pq; struct hfi1_user_sdma_pkt_q *pq = req->pq;
struct sdma_mmu_node *node = NULL; struct sdma_mmu_node *node = NULL;
...@@ -1048,7 +1076,9 @@ static int pin_vector_pages(struct user_sdma_request *req, ...@@ -1048,7 +1076,9 @@ static int pin_vector_pages(struct user_sdma_request *req,
node->rb.addr = (unsigned long)iovec->iov.iov_base; node->rb.addr = (unsigned long)iovec->iov.iov_base;
node->rb.len = iovec->iov.iov_len; node->rb.len = iovec->iov.iov_len;
node->pq = pq;
atomic_set(&node->refcount, 0); atomic_set(&node->refcount, 0);
INIT_LIST_HEAD(&node->list);
} }
npages = num_user_pages(&iovec->iov); npages = num_user_pages(&iovec->iov);
...@@ -1062,6 +1092,14 @@ static int pin_vector_pages(struct user_sdma_request *req, ...@@ -1062,6 +1092,14 @@ static int pin_vector_pages(struct user_sdma_request *req,
memcpy(pages, node->pages, node->npages * sizeof(*pages)); memcpy(pages, node->pages, node->npages * sizeof(*pages));
npages -= node->npages; npages -= node->npages;
retry:
if (!hfi1_can_pin_pages(pq->dd, pq->n_locked, npages)) {
spin_lock(&pq->evict_lock);
cleared = sdma_cache_evict(pq, npages);
spin_unlock(&pq->evict_lock);
if (cleared >= npages)
goto retry;
}
pinned = hfi1_acquire_user_pages( pinned = hfi1_acquire_user_pages(
((unsigned long)iovec->iov.iov_base + ((unsigned long)iovec->iov.iov_base +
(node->npages * PAGE_SIZE)), npages, 0, (node->npages * PAGE_SIZE)), npages, 0,
...@@ -1080,13 +1118,27 @@ static int pin_vector_pages(struct user_sdma_request *req, ...@@ -1080,13 +1118,27 @@ static int pin_vector_pages(struct user_sdma_request *req,
node->pages = pages; node->pages = pages;
node->npages += pinned; node->npages += pinned;
npages = node->npages; npages = node->npages;
spin_lock(&pq->evict_lock);
if (!rb_node)
list_add(&node->list, &pq->evict);
else
list_move(&node->list, &pq->evict);
pq->n_locked += pinned;
spin_unlock(&pq->evict_lock);
} }
iovec->pages = node->pages; iovec->pages = node->pages;
iovec->npages = npages; iovec->npages = npages;
if (!rb_node) { if (!rb_node) {
if (hfi1_mmu_rb_insert(&req->pq->sdma_rb_root, &node->rb)) ret = hfi1_mmu_rb_insert(&req->pq->sdma_rb_root, &node->rb);
if (ret) {
spin_lock(&pq->evict_lock);
list_del(&node->list);
pq->n_locked -= node->npages;
spin_unlock(&pq->evict_lock);
ret = 0;
goto bail; goto bail;
}
} else { } else {
atomic_inc(&node->refcount); atomic_inc(&node->refcount);
} }
...@@ -1503,6 +1555,11 @@ static void sdma_rb_remove(struct rb_root *root, struct mmu_rb_node *mnode, ...@@ -1503,6 +1555,11 @@ static void sdma_rb_remove(struct rb_root *root, struct mmu_rb_node *mnode,
struct sdma_mmu_node *node = struct sdma_mmu_node *node =
container_of(mnode, struct sdma_mmu_node, rb); container_of(mnode, struct sdma_mmu_node, rb);
spin_lock(&node->pq->evict_lock);
list_del(&node->list);
node->pq->n_locked -= node->npages;
spin_unlock(&node->pq->evict_lock);
unpin_vector_pages(notifier ? NULL : current->mm, node->pages, unpin_vector_pages(notifier ? NULL : current->mm, node->pages,
node->npages); node->npages);
/* /*
......
...@@ -68,6 +68,9 @@ struct hfi1_user_sdma_pkt_q { ...@@ -68,6 +68,9 @@ struct hfi1_user_sdma_pkt_q {
wait_queue_head_t wait; wait_queue_head_t wait;
unsigned long unpinned; unsigned long unpinned;
struct rb_root sdma_rb_root; struct rb_root sdma_rb_root;
u32 n_locked;
struct list_head evict;
spinlock_t evict_lock; /* protect evict and n_locked */
}; };
struct hfi1_user_sdma_comp_q { struct hfi1_user_sdma_comp_q {
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment