diff --git a/mm/internal.h b/mm/internal.h index a4b927cdca096ebf55505b2ae4eae5d53f339543..4fe67a162cb4ef022d3555d336fc34e005a0f0b2 100644 --- a/mm/internal.h +++ b/mm/internal.h @@ -63,18 +63,6 @@ static inline unsigned long page_order(struct page *page) return page_private(page); } -/* - * unevictable_migrate_page() called only from migrate_page_copy() to - * migrate unevictable flag to new page. - * Note that the old page has been isolated from the LRU lists at this - * point so we don't need to worry about LRU statistics. - */ -static inline void unevictable_migrate_page(struct page *new, struct page *old) -{ - if (TestClearPageUnevictable(old)) - SetPageUnevictable(new); -} - #ifdef CONFIG_MMU extern long mlock_vma_pages_range(struct vm_area_struct *vma, unsigned long start, unsigned long end); diff --git a/mm/migrate.c b/mm/migrate.c index 2a0ea3ef509eb063118eef9cc2d43b1ddc58ffc1..efddbf0926b283ae5ef292e076ff9900842c4dd5 100644 --- a/mm/migrate.c +++ b/mm/migrate.c @@ -288,8 +288,8 @@ static void migrate_page_copy(struct page *newpage, struct page *page) if (TestClearPageActive(page)) { VM_BUG_ON(PageUnevictable(page)); SetPageActive(newpage); - } else - unevictable_migrate_page(newpage, page); + } else if (TestClearPageUnevictable(page)) + SetPageUnevictable(newpage); if (PageChecked(page)) SetPageChecked(newpage); if (PageMappedToDisk(page))