Revert "BACKPORT: mm: page_isolation: prepare for hygienic freelists"
This reverts commit f15ddfd378
.
Change-Id: I214a4d6460d2a7393a5accacb2ce0c85af5bbaa9
Bug: 420771453
Signed-off-by: Kalesh Singh <kaleshsingh@google.com>
This commit is contained in:
@@ -34,9 +34,7 @@ static inline bool is_migrate_isolate(int migratetype)
|
|||||||
#define REPORT_FAILURE 0x2
|
#define REPORT_FAILURE 0x2
|
||||||
|
|
||||||
void set_pageblock_migratetype(struct page *page, int migratetype);
|
void set_pageblock_migratetype(struct page *page, int migratetype);
|
||||||
|
int move_freepages_block(struct zone *zone, struct page *page, int migratetype);
|
||||||
bool move_freepages_block_isolate(struct zone *zone, struct page *page,
|
|
||||||
int migratetype);
|
|
||||||
|
|
||||||
int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn,
|
int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn,
|
||||||
int migratetype, int flags, gfp_t gfp_flags);
|
int migratetype, int flags, gfp_t gfp_flags);
|
||||||
|
@@ -736,6 +736,10 @@ extern void *memmap_alloc(phys_addr_t size, phys_addr_t align,
|
|||||||
void memmap_init_range(unsigned long, int, unsigned long, unsigned long,
|
void memmap_init_range(unsigned long, int, unsigned long, unsigned long,
|
||||||
unsigned long, enum meminit_context, struct vmem_altmap *, int);
|
unsigned long, enum meminit_context, struct vmem_altmap *, int);
|
||||||
|
|
||||||
|
|
||||||
|
int split_free_page(struct page *free_page,
|
||||||
|
unsigned int order, unsigned long split_pfn_offset);
|
||||||
|
|
||||||
#if defined CONFIG_COMPACTION || defined CONFIG_CMA
|
#if defined CONFIG_COMPACTION || defined CONFIG_CMA
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
204
mm/page_alloc.c
204
mm/page_alloc.c
@@ -961,6 +961,66 @@ done_merging:
|
|||||||
page_reporting_notify_free(order);
|
page_reporting_notify_free(order);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* split_free_page() -- split a free page at split_pfn_offset
|
||||||
|
* @free_page: the original free page
|
||||||
|
* @order: the order of the page
|
||||||
|
* @split_pfn_offset: split offset within the page
|
||||||
|
*
|
||||||
|
* Return -ENOENT if the free page is changed, otherwise 0
|
||||||
|
*
|
||||||
|
* It is used when the free page crosses two pageblocks with different migratetypes
|
||||||
|
* at split_pfn_offset within the page. The split free page will be put into
|
||||||
|
* separate migratetype lists afterwards. Otherwise, the function achieves
|
||||||
|
* nothing.
|
||||||
|
*/
|
||||||
|
int split_free_page(struct page *free_page,
|
||||||
|
unsigned int order, unsigned long split_pfn_offset)
|
||||||
|
{
|
||||||
|
struct zone *zone = page_zone(free_page);
|
||||||
|
unsigned long free_page_pfn = page_to_pfn(free_page);
|
||||||
|
unsigned long pfn;
|
||||||
|
unsigned long flags;
|
||||||
|
int free_page_order;
|
||||||
|
int mt;
|
||||||
|
int ret = 0;
|
||||||
|
|
||||||
|
VM_WARN_ON_ONCE_PAGE(!page_can_split(free_page), free_page);
|
||||||
|
|
||||||
|
if (split_pfn_offset == 0)
|
||||||
|
return ret;
|
||||||
|
|
||||||
|
spin_lock_irqsave(&zone->lock, flags);
|
||||||
|
|
||||||
|
if (!PageBuddy(free_page) || buddy_order(free_page) != order) {
|
||||||
|
ret = -ENOENT;
|
||||||
|
goto out;
|
||||||
|
}
|
||||||
|
|
||||||
|
mt = get_pfnblock_migratetype(free_page, free_page_pfn);
|
||||||
|
if (likely(!is_migrate_isolate(mt)))
|
||||||
|
__mod_zone_freepage_state(zone, -(1UL << order), mt);
|
||||||
|
|
||||||
|
del_page_from_free_list(free_page, zone, order);
|
||||||
|
for (pfn = free_page_pfn;
|
||||||
|
pfn < free_page_pfn + (1UL << order);) {
|
||||||
|
int mt = get_pfnblock_migratetype(pfn_to_page(pfn), pfn);
|
||||||
|
|
||||||
|
free_page_order = min_t(unsigned int,
|
||||||
|
pfn ? __ffs(pfn) : order,
|
||||||
|
__fls(split_pfn_offset));
|
||||||
|
__free_one_page(pfn_to_page(pfn), pfn, zone, free_page_order,
|
||||||
|
mt, FPI_NONE);
|
||||||
|
pfn += 1UL << free_page_order;
|
||||||
|
split_pfn_offset -= (1UL << free_page_order);
|
||||||
|
/* we have done the first part, now switch to second part */
|
||||||
|
if (split_pfn_offset == 0)
|
||||||
|
split_pfn_offset = (1UL << order) - (pfn - free_page_pfn);
|
||||||
|
}
|
||||||
|
out:
|
||||||
|
spin_unlock_irqrestore(&zone->lock, flags);
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
/*
|
/*
|
||||||
* A bad page could be due to a number of fields. Instead of multiple branches,
|
* A bad page could be due to a number of fields. Instead of multiple branches,
|
||||||
* try and check multiple fields with one check. The caller must do a detailed
|
* try and check multiple fields with one check. The caller must do a detailed
|
||||||
@@ -1815,7 +1875,7 @@ static bool prep_move_freepages_block(struct zone *zone, struct page *page,
|
|||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int move_freepages_block(struct zone *zone, struct page *page,
|
int move_freepages_block(struct zone *zone, struct page *page,
|
||||||
int migratetype)
|
int migratetype)
|
||||||
{
|
{
|
||||||
unsigned long start_pfn, end_pfn;
|
unsigned long start_pfn, end_pfn;
|
||||||
@@ -1827,123 +1887,6 @@ static int move_freepages_block(struct zone *zone, struct page *page,
|
|||||||
return move_freepages(zone, start_pfn, end_pfn, migratetype);
|
return move_freepages(zone, start_pfn, end_pfn, migratetype);
|
||||||
}
|
}
|
||||||
|
|
||||||
#ifdef CONFIG_MEMORY_ISOLATION
|
|
||||||
/* Look for a buddy that straddles start_pfn */
|
|
||||||
static unsigned long find_large_buddy(unsigned long start_pfn)
|
|
||||||
{
|
|
||||||
int order = 0;
|
|
||||||
struct page *page;
|
|
||||||
unsigned long pfn = start_pfn;
|
|
||||||
|
|
||||||
while (!PageBuddy(page = pfn_to_page(pfn))) {
|
|
||||||
/* Nothing found */
|
|
||||||
if (++order > MAX_ORDER)
|
|
||||||
return start_pfn;
|
|
||||||
pfn &= ~0UL << order;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Found a preceding buddy, but does it straddle?
|
|
||||||
*/
|
|
||||||
if (pfn + (1 << buddy_order(page)) > start_pfn)
|
|
||||||
return pfn;
|
|
||||||
|
|
||||||
/* Nothing found */
|
|
||||||
return start_pfn;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Split a multi-block free page into its individual pageblocks */
|
|
||||||
static void split_large_buddy(struct zone *zone, struct page *page,
|
|
||||||
unsigned long pfn, int order)
|
|
||||||
{
|
|
||||||
unsigned long end_pfn = pfn + (1 << order);
|
|
||||||
|
|
||||||
VM_WARN_ON_ONCE(order <= pageblock_order);
|
|
||||||
VM_WARN_ON_ONCE(pfn & (pageblock_nr_pages - 1));
|
|
||||||
|
|
||||||
/* Caller removed page from freelist, buddy info cleared! */
|
|
||||||
VM_WARN_ON_ONCE(PageBuddy(page));
|
|
||||||
|
|
||||||
while (pfn != end_pfn) {
|
|
||||||
int mt = get_pfnblock_migratetype(page, pfn);
|
|
||||||
|
|
||||||
__free_one_page(page, pfn, zone, pageblock_order, mt, FPI_NONE);
|
|
||||||
pfn += pageblock_nr_pages;
|
|
||||||
page = pfn_to_page(pfn);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* move_freepages_block_isolate - move free pages in block for page isolation
|
|
||||||
* @zone: the zone
|
|
||||||
* @page: the pageblock page
|
|
||||||
* @migratetype: migratetype to set on the pageblock
|
|
||||||
*
|
|
||||||
* This is similar to move_freepages_block(), but handles the special
|
|
||||||
* case encountered in page isolation, where the block of interest
|
|
||||||
* might be part of a larger buddy spanning multiple pageblocks.
|
|
||||||
*
|
|
||||||
* Unlike the regular page allocator path, which moves pages while
|
|
||||||
* stealing buddies off the freelist, page isolation is interested in
|
|
||||||
* arbitrary pfn ranges that may have overlapping buddies on both ends.
|
|
||||||
*
|
|
||||||
* This function handles that. Straddling buddies are split into
|
|
||||||
* individual pageblocks. Only the block of interest is moved.
|
|
||||||
*
|
|
||||||
* Returns %true if pages could be moved, %false otherwise.
|
|
||||||
*/
|
|
||||||
bool move_freepages_block_isolate(struct zone *zone, struct page *page,
|
|
||||||
int migratetype)
|
|
||||||
{
|
|
||||||
unsigned long start_pfn, end_pfn, pfn;
|
|
||||||
int nr_moved, mt;
|
|
||||||
|
|
||||||
if (!prep_move_freepages_block(zone, page, &start_pfn, &end_pfn,
|
|
||||||
NULL, NULL))
|
|
||||||
return false;
|
|
||||||
|
|
||||||
/* No splits needed if buddies can't span multiple blocks */
|
|
||||||
if (pageblock_order == MAX_ORDER)
|
|
||||||
goto move;
|
|
||||||
|
|
||||||
/* We're a tail block in a larger buddy */
|
|
||||||
pfn = find_large_buddy(start_pfn);
|
|
||||||
if (pfn != start_pfn) {
|
|
||||||
struct page *buddy = pfn_to_page(pfn);
|
|
||||||
int order = buddy_order(buddy);
|
|
||||||
int mt = get_pfnblock_migratetype(buddy, pfn);
|
|
||||||
|
|
||||||
if (!is_migrate_isolate(mt))
|
|
||||||
__mod_zone_freepage_state(zone, -(1UL << order), mt);
|
|
||||||
del_page_from_free_list(buddy, zone, order);
|
|
||||||
set_pageblock_migratetype(page, migratetype);
|
|
||||||
split_large_buddy(zone, buddy, pfn, order);
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* We're the starting block of a larger buddy */
|
|
||||||
if (PageBuddy(page) && buddy_order(page) > pageblock_order) {
|
|
||||||
int mt = get_pfnblock_migratetype(page, pfn);
|
|
||||||
int order = buddy_order(page);
|
|
||||||
|
|
||||||
if (!is_migrate_isolate(mt))
|
|
||||||
__mod_zone_freepage_state(zone, -(1UL << order), mt);
|
|
||||||
del_page_from_free_list(page, zone, order);
|
|
||||||
set_pageblock_migratetype(page, migratetype);
|
|
||||||
split_large_buddy(zone, page, pfn, order);
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
move:
|
|
||||||
mt = get_pfnblock_migratetype(page, start_pfn);
|
|
||||||
nr_moved = move_freepages(zone, start_pfn, end_pfn, migratetype);
|
|
||||||
if (!is_migrate_isolate(mt))
|
|
||||||
__mod_zone_freepage_state(zone, -nr_moved, mt);
|
|
||||||
else if (!is_migrate_isolate(migratetype))
|
|
||||||
__mod_zone_freepage_state(zone, nr_moved, migratetype);
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
#endif /* CONFIG_MEMORY_ISOLATION */
|
|
||||||
|
|
||||||
static void change_pageblock_range(struct page *pageblock_page,
|
static void change_pageblock_range(struct page *pageblock_page,
|
||||||
int start_order, int migratetype)
|
int start_order, int migratetype)
|
||||||
{
|
{
|
||||||
@@ -6617,6 +6560,7 @@ int alloc_contig_range(unsigned long start, unsigned long end,
|
|||||||
unsigned migratetype, gfp_t gfp_mask)
|
unsigned migratetype, gfp_t gfp_mask)
|
||||||
{
|
{
|
||||||
unsigned long outer_start, outer_end;
|
unsigned long outer_start, outer_end;
|
||||||
|
int order;
|
||||||
int ret = 0;
|
int ret = 0;
|
||||||
|
|
||||||
struct compact_control cc = {
|
struct compact_control cc = {
|
||||||
@@ -6693,7 +6637,29 @@ int alloc_contig_range(unsigned long start, unsigned long end,
|
|||||||
* We don't have to hold zone->lock here because the pages are
|
* We don't have to hold zone->lock here because the pages are
|
||||||
* isolated thus they won't get removed from buddy.
|
* isolated thus they won't get removed from buddy.
|
||||||
*/
|
*/
|
||||||
outer_start = find_large_buddy(start);
|
|
||||||
|
order = 0;
|
||||||
|
outer_start = start;
|
||||||
|
while (!PageBuddy(pfn_to_page(outer_start))) {
|
||||||
|
if (++order > MAX_ORDER) {
|
||||||
|
outer_start = start;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
outer_start &= ~0UL << order;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (outer_start != start) {
|
||||||
|
order = buddy_order(pfn_to_page(outer_start));
|
||||||
|
|
||||||
|
/*
|
||||||
|
* outer_start page could be small order buddy page and
|
||||||
|
* it doesn't include start page. Adjust outer_start
|
||||||
|
* in this case to report failed page properly
|
||||||
|
* on tracepoint in test_pages_isolated()
|
||||||
|
*/
|
||||||
|
if (outer_start + (1UL << order) <= start)
|
||||||
|
outer_start = start;
|
||||||
|
}
|
||||||
|
|
||||||
/* Make sure the range is really isolated. */
|
/* Make sure the range is really isolated. */
|
||||||
if (test_pages_isolated(outer_start, end, 0)) {
|
if (test_pages_isolated(outer_start, end, 0)) {
|
||||||
|
@@ -179,10 +179,16 @@ static int set_migratetype_isolate(struct page *page, int migratetype, int isol_
|
|||||||
unmovable = has_unmovable_pages(check_unmovable_start, check_unmovable_end,
|
unmovable = has_unmovable_pages(check_unmovable_start, check_unmovable_end,
|
||||||
migratetype, isol_flags);
|
migratetype, isol_flags);
|
||||||
if (!unmovable) {
|
if (!unmovable) {
|
||||||
if (!move_freepages_block_isolate(zone, page, MIGRATE_ISOLATE)) {
|
int nr_pages;
|
||||||
|
int mt = get_pageblock_migratetype(page);
|
||||||
|
|
||||||
|
nr_pages = move_freepages_block(zone, page, MIGRATE_ISOLATE);
|
||||||
|
/* Block spans zone boundaries? */
|
||||||
|
if (nr_pages == -1) {
|
||||||
spin_unlock_irqrestore(&zone->lock, flags);
|
spin_unlock_irqrestore(&zone->lock, flags);
|
||||||
return -EBUSY;
|
return -EBUSY;
|
||||||
}
|
}
|
||||||
|
__mod_zone_freepage_state(zone, -nr_pages, mt);
|
||||||
zone->nr_isolate_pageblock++;
|
zone->nr_isolate_pageblock++;
|
||||||
spin_unlock_irqrestore(&zone->lock, flags);
|
spin_unlock_irqrestore(&zone->lock, flags);
|
||||||
return 0;
|
return 0;
|
||||||
@@ -249,11 +255,13 @@ static void unset_migratetype_isolate(struct page *page, int migratetype)
|
|||||||
* allocation.
|
* allocation.
|
||||||
*/
|
*/
|
||||||
if (!isolated_page) {
|
if (!isolated_page) {
|
||||||
|
int nr_pages = move_freepages_block(zone, page, migratetype);
|
||||||
/*
|
/*
|
||||||
* Isolating this block already succeeded, so this
|
* Isolating this block already succeeded, so this
|
||||||
* should not fail on zone boundaries.
|
* should not fail on zone boundaries.
|
||||||
*/
|
*/
|
||||||
WARN_ON_ONCE(!move_freepages_block_isolate(zone, page, migratetype));
|
WARN_ON_ONCE(nr_pages == -1);
|
||||||
|
__mod_zone_freepage_state(zone, nr_pages, migratetype);
|
||||||
} else {
|
} else {
|
||||||
set_pageblock_migratetype(page, migratetype);
|
set_pageblock_migratetype(page, migratetype);
|
||||||
__putback_isolated_page(page, order, migratetype);
|
__putback_isolated_page(page, order, migratetype);
|
||||||
@@ -366,29 +374,26 @@ static int isolate_single_pageblock(unsigned long boundary_pfn, int flags,
|
|||||||
|
|
||||||
VM_BUG_ON(!page);
|
VM_BUG_ON(!page);
|
||||||
pfn = page_to_pfn(page);
|
pfn = page_to_pfn(page);
|
||||||
|
/*
|
||||||
|
* start_pfn is MAX_ORDER_NR_PAGES aligned, if there is any
|
||||||
|
* free pages in [start_pfn, boundary_pfn), its head page will
|
||||||
|
* always be in the range.
|
||||||
|
*/
|
||||||
if (PageBuddy(page)) {
|
if (PageBuddy(page)) {
|
||||||
int order = buddy_order(page);
|
int order = buddy_order(page);
|
||||||
|
|
||||||
/* move_freepages_block_isolate() handled this */
|
if (pfn + (1UL << order) > boundary_pfn) {
|
||||||
VM_WARN_ON_ONCE(pfn + (1 << order) > boundary_pfn);
|
/* free page changed before split, check it again */
|
||||||
|
if (split_free_page(page, order, boundary_pfn - pfn))
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
pfn += 1UL << order;
|
pfn += 1UL << order;
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* If a compound page is straddling our block, attempt
|
* migrate compound pages then let the free page handling code
|
||||||
* to migrate it out of the way.
|
* above do the rest. If migration is not possible, just fail.
|
||||||
*
|
|
||||||
* We don't have to worry about this creating a large
|
|
||||||
* free page that straddles into our block: gigantic
|
|
||||||
* pages are freed as order-0 chunks, and LRU pages
|
|
||||||
* (currently) do not exceed pageblock_order.
|
|
||||||
*
|
|
||||||
* The block of interest has already been marked
|
|
||||||
* MIGRATE_ISOLATE above, so when migration is done it
|
|
||||||
* will free its pages onto the correct freelists.
|
|
||||||
*/
|
*/
|
||||||
if (PageCompound(page)) {
|
if (PageCompound(page)) {
|
||||||
struct page *head = compound_head(page);
|
struct page *head = compound_head(page);
|
||||||
@@ -399,10 +404,16 @@ static int isolate_single_pageblock(unsigned long boundary_pfn, int flags,
|
|||||||
pfn = head_pfn + nr_pages;
|
pfn = head_pfn + nr_pages;
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
#if defined CONFIG_COMPACTION || defined CONFIG_CMA
|
#if defined CONFIG_COMPACTION || defined CONFIG_CMA
|
||||||
if (PageHuge(page)) {
|
/*
|
||||||
|
* hugetlb, lru compound (THP), and movable compound pages
|
||||||
|
* can be migrated. Otherwise, fail the isolation.
|
||||||
|
*/
|
||||||
|
if (PageHuge(page) || PageLRU(page) || __PageMovable(page)) {
|
||||||
|
int order;
|
||||||
|
unsigned long outer_pfn;
|
||||||
int page_mt = get_pageblock_migratetype(page);
|
int page_mt = get_pageblock_migratetype(page);
|
||||||
|
bool isolate_page = !is_migrate_isolate_page(page);
|
||||||
struct compact_control cc = {
|
struct compact_control cc = {
|
||||||
.nr_migratepages = 0,
|
.nr_migratepages = 0,
|
||||||
.order = -1,
|
.order = -1,
|
||||||
@@ -415,24 +426,54 @@ static int isolate_single_pageblock(unsigned long boundary_pfn, int flags,
|
|||||||
};
|
};
|
||||||
INIT_LIST_HEAD(&cc.migratepages);
|
INIT_LIST_HEAD(&cc.migratepages);
|
||||||
|
|
||||||
ret = __alloc_contig_migrate_range(&cc, head_pfn,
|
/*
|
||||||
head_pfn + nr_pages, page_mt);
|
* XXX: mark the page as MIGRATE_ISOLATE so that
|
||||||
|
* no one else can grab the freed page after migration.
|
||||||
|
* Ideally, the page should be freed as two separate
|
||||||
|
* pages to be added into separate migratetype free
|
||||||
|
* lists.
|
||||||
|
*/
|
||||||
|
if (isolate_page) {
|
||||||
|
ret = set_migratetype_isolate(page, page_mt,
|
||||||
|
flags, head_pfn, head_pfn + nr_pages);
|
||||||
if (ret)
|
if (ret)
|
||||||
goto failed;
|
goto failed;
|
||||||
|
|
||||||
pfn = head_pfn + nr_pages;
|
|
||||||
continue;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
ret = __alloc_contig_migrate_range(&cc, head_pfn,
|
||||||
|
head_pfn + nr_pages, page_mt);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* These pages are movable too, but they're
|
* restore the page's migratetype so that it can
|
||||||
* not expected to exceed pageblock_order.
|
* be split into separate migratetype free lists
|
||||||
*
|
* later.
|
||||||
* Let us know when they do, so we can add
|
|
||||||
* proper free and split handling for them.
|
|
||||||
*/
|
*/
|
||||||
VM_WARN_ON_ONCE_PAGE(PageLRU(page), page);
|
if (isolate_page)
|
||||||
VM_WARN_ON_ONCE_PAGE(__PageMovable(page), page);
|
unset_migratetype_isolate(page, page_mt);
|
||||||
|
|
||||||
|
if (ret)
|
||||||
|
goto failed;
|
||||||
|
/*
|
||||||
|
* reset pfn to the head of the free page, so
|
||||||
|
* that the free page handling code above can split
|
||||||
|
* the free page to the right migratetype list.
|
||||||
|
*
|
||||||
|
* head_pfn is not used here as a hugetlb page order
|
||||||
|
* can be bigger than MAX_ORDER, but after it is
|
||||||
|
* freed, the free page order is not. Use pfn within
|
||||||
|
* the range to find the head of the free page.
|
||||||
|
*/
|
||||||
|
order = 0;
|
||||||
|
outer_pfn = pfn;
|
||||||
|
while (!PageBuddy(pfn_to_page(outer_pfn))) {
|
||||||
|
/* stop if we cannot find the free page */
|
||||||
|
if (++order > MAX_ORDER)
|
||||||
|
goto failed;
|
||||||
|
outer_pfn &= ~0UL << order;
|
||||||
|
}
|
||||||
|
pfn = outer_pfn;
|
||||||
|
continue;
|
||||||
|
} else
|
||||||
#endif
|
#endif
|
||||||
goto failed;
|
goto failed;
|
||||||
}
|
}
|
||||||
|
Reference in New Issue
Block a user