From patchwork Mon Jul 6 17:34:24 2015 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Grygorii Strashko X-Patchwork-Id: 6726941 X-Patchwork-Delegate: rjw@sisk.pl Return-Path: X-Original-To: patchwork-linux-pm@patchwork.kernel.org Delivered-To: patchwork-parsemail@patchwork1.web.kernel.org Received: from mail.kernel.org (mail.kernel.org [198.145.29.136]) by patchwork1.web.kernel.org (Postfix) with ESMTP id 513BA9F380 for ; Mon, 6 Jul 2015 17:35:10 +0000 (UTC) Received: from mail.kernel.org (localhost [127.0.0.1]) by mail.kernel.org (Postfix) with ESMTP id 9242520694 for ; Mon, 6 Jul 2015 17:35:08 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id 95F20205FC for ; Mon, 6 Jul 2015 17:35:06 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752876AbbGFRfF (ORCPT ); Mon, 6 Jul 2015 13:35:05 -0400 Received: from arroyo.ext.ti.com ([192.94.94.40]:57611 "EHLO arroyo.ext.ti.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751164AbbGFRfA (ORCPT ); Mon, 6 Jul 2015 13:35:00 -0400 Received: from dlelxv90.itg.ti.com ([172.17.2.17]) by arroyo.ext.ti.com (8.13.7/8.13.7) with ESMTP id t66HYSwJ025550; Mon, 6 Jul 2015 12:34:28 -0500 Received: from DFLE72.ent.ti.com (dfle72.ent.ti.com [128.247.5.109]) by dlelxv90.itg.ti.com (8.14.3/8.13.8) with ESMTP id t66HYROI021559; Mon, 6 Jul 2015 12:34:27 -0500 Received: from dlep33.itg.ti.com (157.170.170.75) by DFLE72.ent.ti.com (128.247.5.109) with Microsoft SMTP Server id 14.3.224.2; Mon, 6 Jul 2015 12:34:15 -0500 Received: from localhost (ileax41-snat.itg.ti.com [10.172.224.153]) by dlep33.itg.ti.com (8.14.3/8.13.8) with ESMTP id t66HYQv0021745; Mon, 6 Jul 2015 12:34:27 -0500 From: Grygorii Strashko To: "Rafael J. Wysocki" , , Len Brown , Pavel Machek CC: , , , Grygorii Strashko Subject: [PATCH v2] PM / hibernate: fix kernel crash in alloc_highmem_pages Date: Mon, 6 Jul 2015 20:34:24 +0300 Message-ID: <1436204064-22925-1-git-send-email-grygorii.strashko@ti.com> X-Mailer: git-send-email 2.4.5 MIME-Version: 1.0 Sender: linux-pm-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-pm@vger.kernel.org X-Spam-Status: No, score=-7.6 required=5.0 tests=BAYES_00, RCVD_IN_DNSWL_HI, RP_MATCHES_RCVD, UNPARSEABLE_RELAY autolearn=unavailable version=3.3.1 X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on mail.kernel.org X-Virus-Scanned: ClamAV using ClamSMTP Now Kernel crashes under some circumstances in swsusp_alloc() because it hits BUG() in memory_bm_set_bit(). Kernel: Linux 4.1-rc3 HW: TI dra7-evm board, 2xARM Cortex-A15, 1.5G RAM Swap: 2G swap partition on SD-card Steps to reproduce: - reserve 500M memory for CMA from command line: "cma=500M" - configure image_size # echo 1073741824 > sys/power/image_size - enable swap # swapon /dev/ - run memtester # memtester 460M & - [optional] disable console suspend # echo 0 > sys/module/printk/parameters/console_suspend - perform suspend to disk # echo disk > sys/power/state Crash report: -----------[ cut here ]------------ Kernel BUG at c0097394 [verbose debug info unavailable] Internal error: Oops - BUG: 0 [#1] SMP ARM Modules linked in: CPU: 0 PID: 68 Comm: sh Not tainted 4.1.0-rc3-00190-ga96463d-dirty #138 Hardware name: Generic DRA74X (Flattened Device Tree) task: ee3b2fc0 ti: ee6e8000 task.ti: ee6e8000 PC is at memory_bm_set_bit+0x30/0x34 LR is at 0x80000 pc : [] lr : [<00080000>] psr: a00000d3 sp : ee6e9de4 ip : c111f714 fp : 00000000 r10: 000128e2 r9 : c111f6f8 r8 : 00028658 r7 : c114ef00 r6 : c08bad04 r5 : 00027849 r4 : 00012b59 r3 : ee6e9de8 r2 : ee6e9de4 r1 : 8ebc3c00 r0 : fffffff2 Flags: NzCv IRQs off FIQs off Mode SVC_32 ISA ARM Segment user Control: 10c5387d Table: ae6dc06a DAC: 00000015 Process sh (pid: 68, stack limit = 0xee6e8218) Stack: (0xee6e9de4 to 0xee6ea000) 9de0: 00000000 c0097690 00012b59 c0099148 00000030 ee6e9e4c 00000000 9e00: 00000007 0000000f c096b624 c08b78d4 c001452c 00000030 c00141b0 ae0d0000 9e20: ee6e9e4c 80021ae0 00000000 001fc4c0 001fc4c0 00000015 8000406a 10c5387d 9e40: 00000040 00f00000 00000000 00000000 ee6f4c00 00000000 c08b6800 c0773a60 9e60: c096b624 c096dbe8 00000000 c001435c ee3b2fc0 00000000 c096b624 c00966a8 9e80: c08f9710 00000000 c111f5c8 00200660 c08f9644 00000005 00000000 c0096ef0 9ea0: ee6470c0 00000005 00000004 00200660 ee6a6300 c00944cc 00000005 ee6470c0 9ec0: 00000005 ee6a6300 00200660 ee6470cc ee6e9f88 c01dc688 00000000 00000000 9ee0: 20000013 ee6f76c0 00000005 00200660 ee6e9f88 c000f6a4 ee6e8000 00000000 9f00: 00000002 c016beec c08b87bc c016e32c 00000020 00000000 c016c830 ee3b2fb8 9f20: c016c830 ee652ae4 ee6f76c0 00000000 00000000 c016c51c 00000000 00000003 9f40: 00000000 ee6f76c0 00000005 00200660 ee6f76c0 00000005 00200660 ee6e9f88 9f60: c000f6a4 c016c784 00000000 00000000 ee6f76c0 ee6f76c0 00200660 00000005 9f80: c000f6a4 c016cfb0 00000000 00000000 00200660 001fc4e0 00000005 00200660 9fa0: 00000004 c000f4e0 001fc4e0 00000005 00000001 00200660 00000005 001fbc18 9fc0: 001fc4e0 00000005 00200660 00000004 001fd264 001fcfc8 00000000 00000002 9fe0: 00000001 bee87618 0000c55c 0000921c 60000010 00000001 00000000 00000000 [] (memory_bm_set_bit) from [] (swsusp_save+0x3f0/0x440) [] (swsusp_save) from [] (arch_save_image+0x8/0x2c) [] (arch_save_image) from [] (cpu_suspend_abort+0x0/0x30) [] (cpu_suspend_abort) from [<00000000>] ( (null)) Code: e59d0004 eb0ae150 e28dd00c e49df004 (e7f001f2) ---[ end trace 61e3b86f908e4d7f ]--- As was investigated, the issue depends on following things: 1) Hibernation should follow through the below code path in hibernate_preallocate_memory(): int hibernate_preallocate_memory(void) { [...] /* * If the desired number of image pages is at least as large as the * current number of saveable pages in memory, allocate page frames for * the image and we're done. */ if (size >= saveable) { pages = preallocate_image_highmem(save_highmem); pages += preallocate_image_memory(saveable - pages, avail_normal); goto out; } 2) Amount of memory reserved for CMA purposes. CMA memory reported as free by MM, but Hibernation core can't use it. As result, All memory checks are passed and system reaches later Hibernation stages, tries to create Hibernation immage and, finally crashes in swsusp_alloc(): swsusp_alloc() |- alloc_highmem_pages() |- alloc_image_page(__GFP_HIGHMEM) -- >"sh: page allocation failure: order:0, mode:0x2" |- memory_bm_set_bit |- BUG_ON(error) 3) alloc_page(__GFP_HIGHMEM) doesn't return Highmem pages always, instead MM allocates pages from all available zones evenly. For example: saveable pages 167191 = hmem 80400 + normal 86791 request hmem 80400 - alloc hmem 204 - alloc normal 161 - alloc hmem 442 ... - alloc normal 38615 - alloc hmem 41785 Such behaviour is not always handled properly. As result, swsusp_alloc(), (from above example) will try to allocate (80400 - 41785) additional Highmem pages without taking into account that those pages were allocated already, but from Normal zone. In this patch it is proposed to fix issue by rewriting swsusp_alloc() in the following way: - allocate Highmem buffer - check if we need to allocate any memory at all alloc_normal + alloc_highmem < nr_normal + nr_highmem - calculate number of pages which still need to be allocated nr_pages = nr_normal + nr_highmem - alloc_normal - alloc_highmem - try to allocate Highmem first, but no more than nr_pages - try to get rest from Normal memory - abort if not all pages were allocated And also use preallocate_image_pages() instead of alloc_highmem_pages() and alloc_image_page(). This way, allocated pages will be accounted properly in alloc_highmem and alloc_normal counters. alloc_highmem_pages() isn't used any more after this patch and so removed. Signed-off-by: Grygorii Strashko --- Changes in v2: - applied comments from Rafael. Link on rfc v1: - https://lkml.org/lkml/2015/6/8/406 kernel/power/snapshot.c | 82 ++++++++++++++++++++++--------------------------- 1 file changed, 36 insertions(+), 46 deletions(-) diff --git a/kernel/power/snapshot.c b/kernel/power/snapshot.c index 5235dd4..7b96272 100644 --- a/kernel/power/snapshot.c +++ b/kernel/power/snapshot.c @@ -1761,34 +1761,9 @@ static inline int get_highmem_buffer(int safe_needed) return buffer ? 0 : -ENOMEM; } -/** - * alloc_highmem_image_pages - allocate some highmem pages for the image. - * Try to allocate as many pages as needed, but if the number of free - * highmem pages is lesser than that, allocate them all. - */ - -static inline unsigned int -alloc_highmem_pages(struct memory_bitmap *bm, unsigned int nr_highmem) -{ - unsigned int to_alloc = count_free_highmem_pages(); - - if (to_alloc > nr_highmem) - to_alloc = nr_highmem; - - nr_highmem -= to_alloc; - while (to_alloc-- > 0) { - struct page *page; - - page = alloc_image_page(__GFP_HIGHMEM); - memory_bm_set_bit(bm, page_to_pfn(page)); - } - return nr_highmem; -} #else static inline int get_highmem_buffer(int safe_needed) { return 0; } -static inline unsigned int -alloc_highmem_pages(struct memory_bitmap *bm, unsigned int n) { return 0; } #endif /* CONFIG_HIGHMEM */ /** @@ -1805,29 +1780,44 @@ alloc_highmem_pages(struct memory_bitmap *bm, unsigned int n) { return 0; } static int swsusp_alloc(struct memory_bitmap *orig_bm, struct memory_bitmap *copy_bm, - unsigned int nr_pages, unsigned int nr_highmem) + unsigned int nr_normal, unsigned int nr_highmem) { - if (nr_highmem > 0) { - if (get_highmem_buffer(PG_ANY)) - goto err_out; - if (nr_highmem > alloc_highmem) { - nr_highmem -= alloc_highmem; - nr_pages += alloc_highmem_pages(copy_bm, nr_highmem); - } - } - if (nr_pages > alloc_normal) { - nr_pages -= alloc_normal; - while (nr_pages-- > 0) { - struct page *page; - - page = alloc_image_page(GFP_ATOMIC | __GFP_COLD); - if (!page) - goto err_out; - memory_bm_set_bit(copy_bm, page_to_pfn(page)); - } - } + unsigned int nr_pages; + /* + * Additional input data: + * alloc_highmem - number of allocated Highmem pages + * alloc_normal - number of allocated Normal pages + */ - return 0; + /* allocate Highmem buffer */ + if ((nr_highmem > 0) && get_highmem_buffer(PG_ANY)) + goto err_out; + + /* + * check if we need to allocate any memory at all + * alloc_normal + alloc_highmem < nr_normal + nr_highmem + * and calculate number of pages which still need to be allocated + * nr_pages = nr_normal + nr_highmem - alloc_normal - alloc_highmem + */ + nr_pages = nr_normal + nr_highmem; + if (nr_pages > alloc_normal + alloc_highmem) + nr_pages -= alloc_normal + alloc_highmem; + else + return 0; + + /* try to allocate Highmem first, but no more than nr_pages */ + if (nr_highmem > 0) + nr_pages -= preallocate_image_pages(nr_pages, __GFP_HIGHMEM); + + /* try to get rest from Normal memory */ + if (nr_pages) + nr_pages -= preallocate_image_pages(nr_pages, + GFP_ATOMIC | __GFP_COLD); + + if (!nr_pages) + return 0; + + /* abort if not all pages were allocated */ err_out: swsusp_free();