@@ -31,6 +31,10 @@ static DEFINE_XARRAY(sgx_epc_address_space);
* with sgx_global_lru.lock acquired.
*/
static struct sgx_epc_lru_lists sgx_global_lru;
+static inline struct sgx_epc_lru_lists *sgx_lru_lists(struct sgx_epc_page *epc_page)
+{
+ return &sgx_global_lru;
+}
static atomic_long_t sgx_nr_free_pages = ATOMIC_LONG_INIT(0);
@@ -278,6 +282,7 @@ static void sgx_reclaimer_write(struct sgx_epc_page *epc_page,
/**
* sgx_reclaim_pages() - Reclaim EPC pages from the consumers
* @nr_to_scan: Number of EPC pages to scan for reclaim
+ * @ignore_age: Reclaim a page even if it is young
*
* Take a fixed number of pages from the head of the active page pool and
* reclaim them to the enclave's private shmem files. Skip the pages, which have
@@ -291,11 +296,12 @@ static void sgx_reclaimer_write(struct sgx_epc_page *epc_page,
* problematic as it would increase the lock contention too much, which would
* halt forward progress.
*/
-static int __sgx_reclaim_pages(int nr_to_scan)
+static int __sgx_reclaim_pages(int nr_to_scan, bool ignore_age)
{
struct sgx_backing backing[SGX_MAX_NR_TO_RECLAIM];
struct sgx_epc_page *epc_page, *tmp;
struct sgx_encl_page *encl_page;
+ struct sgx_epc_lru_lists *lru;
pgoff_t page_index;
LIST_HEAD(iso);
int ret;
@@ -332,7 +338,8 @@ static int __sgx_reclaim_pages(int nr_to_scan)
list_for_each_entry_safe(epc_page, tmp, &iso, list) {
encl_page = epc_page->encl_owner;
- if (i == SGX_MAX_NR_TO_RECLAIM || !sgx_reclaimer_age(epc_page))
+ if (i == SGX_MAX_NR_TO_RECLAIM ||
+ (!ignore_age && !sgx_reclaimer_age(epc_page)))
goto skip;
page_index = PFN_DOWN(encl_page->desc - encl_page->encl->base);
@@ -350,10 +357,11 @@ static int __sgx_reclaim_pages(int nr_to_scan)
continue;
skip:
- spin_lock(&sgx_global_lru.lock);
+ lru = sgx_lru_lists(epc_page);
+ spin_lock(&lru->lock);
epc_page->flags &= ~SGX_EPC_PAGE_RECLAIM_IN_PROGRESS;
- sgx_epc_move_reclaimable(&sgx_global_lru, epc_page);
- spin_unlock(&sgx_global_lru.lock);
+ sgx_epc_move_reclaimable(lru, epc_page);
+ spin_unlock(&lru->lock);
kref_put(&encl_page->encl->refcount, sgx_encl_release);
}
@@ -377,11 +385,11 @@ static int __sgx_reclaim_pages(int nr_to_scan)
return i;
}
-static int sgx_reclaim_pages(int nr_to_scan)
+static int sgx_reclaim_pages(int nr_to_scan, bool ignore_age)
{
int ret;
- ret = __sgx_reclaim_pages(nr_to_scan);
+ ret = __sgx_reclaim_pages(nr_to_scan, ignore_age);
cond_resched();
return ret;
}
@@ -402,7 +410,7 @@ static bool sgx_should_reclaim(unsigned long watermark)
void sgx_reclaim_direct(void)
{
if (sgx_should_reclaim(SGX_NR_LOW_PAGES))
- __sgx_reclaim_pages(SGX_NR_TO_SCAN);
+ __sgx_reclaim_pages(SGX_NR_TO_SCAN, false);
}
static int ksgxd(void *p)
@@ -428,7 +436,7 @@ static int ksgxd(void *p)
sgx_should_reclaim(SGX_NR_HIGH_PAGES));
if (sgx_should_reclaim(SGX_NR_HIGH_PAGES))
- sgx_reclaim_pages(SGX_NR_TO_SCAN);
+ sgx_reclaim_pages(SGX_NR_TO_SCAN, false);
}
return 0;
@@ -522,15 +530,17 @@ struct sgx_epc_page *__sgx_alloc_epc_page(void)
*/
void sgx_record_epc_page(struct sgx_epc_page *page, unsigned long flags)
{
- spin_lock(&sgx_global_lru.lock);
+ struct sgx_epc_lru_lists *lru = sgx_lru_lists(page);
+
+ spin_lock(&lru->lock);
WARN_ON(page->flags & (SGX_EPC_PAGE_RECLAIMER_TRACKED |
SGX_EPC_PAGE_RECLAIM_IN_PROGRESS));
page->flags |= flags;
if (flags & SGX_EPC_PAGE_RECLAIMER_TRACKED)
- sgx_epc_push_reclaimable(&sgx_global_lru, page);
+ sgx_epc_push_reclaimable(lru, page);
else
- sgx_epc_push_unreclaimable(&sgx_global_lru, page);
- spin_unlock(&sgx_global_lru.lock);
+ sgx_epc_push_unreclaimable(lru, page);
+ spin_unlock(&lru->lock);
}
/**
@@ -545,14 +555,16 @@ void sgx_record_epc_page(struct sgx_epc_page *page, unsigned long flags)
*/
int sgx_drop_epc_page(struct sgx_epc_page *page)
{
- spin_lock(&sgx_global_lru.lock);
+ struct sgx_epc_lru_lists *lru = sgx_lru_lists(page);
+
+ spin_lock(&lru->lock);
if ((page->flags & SGX_EPC_PAGE_RECLAIMER_TRACKED) &&
(page->flags & SGX_EPC_PAGE_RECLAIM_IN_PROGRESS)) {
- spin_unlock(&sgx_global_lru.lock);
+ spin_unlock(&lru->lock);
return -EBUSY;
}
list_del(&page->list);
- spin_unlock(&sgx_global_lru.lock);
+ spin_unlock(&lru->lock);
page->flags &= ~(SGX_EPC_PAGE_RECLAIMER_TRACKED |
SGX_EPC_PAGE_RECLAIM_IN_PROGRESS |
@@ -607,7 +619,7 @@ struct sgx_epc_page *sgx_alloc_epc_page(void *owner, bool reclaim)
break;
}
- sgx_reclaim_pages(SGX_NR_TO_SCAN);
+ sgx_reclaim_pages(SGX_NR_TO_SCAN, false);
}
if (sgx_should_reclaim(SGX_NR_LOW_PAGES))