From mboxrd@z Thu Jan 1 00:00:00 1970 From: Yinghai Lu Subject: [PATCH 01/39] swiotlb: Use page alignment for early buffer allocation Date: Thu, 8 Apr 2010 23:03:30 -0700 Message-ID: <1270793048-23796-2-git-send-email-yinghai@kernel.org> References: <1270793048-23796-1-git-send-email-yinghai@kernel.org> Return-path: In-Reply-To: <1270793048-23796-1-git-send-email-yinghai@kernel.org> Sender: linux-kernel-owner@vger.kernel.org To: Ingo Molnar , Thomas Gleixner , "H. Peter Anvin" , Andrew Morton Cc: David Miller , Benjamin Herrenschmidt , Linus Torvalds , Johannes Weiner , linux-kernel@vger.kernel.org, linux-arch@vger.kernel.org, Yinghai Lu , FUJITA Tomonori , Becky Bruce List-Id: linux-arch.vger.kernel.org We could call free_bootmem_late() if swiotlb is not used, and it will shrink to page alignement. So alloc them with page alignment at first, to avoid lose two pages before patch: [ 0.000000] lmb_reserve_area: [00d3600000, 00d7600000] swiotlb buffer [ 0.000000] lmb_reserve_area: [00d7e7ef40, 00d7e9ef40] swiotlb list [ 0.000000] lmb_reserve_area: [00d7e3ef40, 00d7e7ef40] swiotlb orig_ad [ 0.000000] lmb_reserve_area: [000008a000, 0000092000] swiotlb overflo after patch will get [ 0.000000] lmb_reserve_area: [00d3600000, 00d7600000] swiotlb buffer [ 0.000000] lmb_reserve_area: [00d7e7e000, 00d7e9e000] swiotlb list [ 0.000000] lmb_reserve_area: [00d7e3e000, 00d7e7e000] swiotlb orig_ad [ 0.000000] lmb_reserve_area: [000008a000, 0000092000] swiotlb overflo Signed-off-by: Yinghai Lu Cc: FUJITA Tomonori Cc: Becky Bruce --- lib/swiotlb.c | 16 ++++++++-------- 1 files changed, 8 insertions(+), 8 deletions(-) diff --git a/lib/swiotlb.c b/lib/swiotlb.c index 5fddf72..1bd4258 100644 --- a/lib/swiotlb.c +++ b/lib/swiotlb.c @@ -159,7 +159,7 @@ swiotlb_init_with_default_size(size_t default_size, int verbose) /* * Get IO TLB memory from the low pages */ - io_tlb_start = alloc_bootmem_low_pages(bytes); + io_tlb_start = alloc_bootmem_low_pages(PAGE_ALIGN(bytes)); if (!io_tlb_start) panic("Cannot allocate SWIOTLB buffer"); io_tlb_end = io_tlb_start + bytes; @@ -169,16 +169,16 @@ swiotlb_init_with_default_size(size_t default_size, int verbose) * to find contiguous free memory regions of size up to IO_TLB_SEGSIZE * between io_tlb_start and io_tlb_end. */ - io_tlb_list = alloc_bootmem(io_tlb_nslabs * sizeof(int)); + io_tlb_list = alloc_bootmem_pages(PAGE_ALIGN(io_tlb_nslabs * sizeof(int))); for (i = 0; i < io_tlb_nslabs; i++) io_tlb_list[i] = IO_TLB_SEGSIZE - OFFSET(i, IO_TLB_SEGSIZE); io_tlb_index = 0; - io_tlb_orig_addr = alloc_bootmem(io_tlb_nslabs * sizeof(phys_addr_t)); + io_tlb_orig_addr = alloc_bootmem_pages(PAGE_ALIGN(io_tlb_nslabs * sizeof(phys_addr_t))); /* * Get the overflow emergency buffer */ - io_tlb_overflow_buffer = alloc_bootmem_low(io_tlb_overflow); + io_tlb_overflow_buffer = alloc_bootmem_low_pages(PAGE_ALIGN(io_tlb_overflow)); if (!io_tlb_overflow_buffer) panic("Cannot allocate SWIOTLB overflow buffer!\n"); if (verbose) @@ -304,13 +304,13 @@ void __init swiotlb_free(void) get_order(io_tlb_nslabs << IO_TLB_SHIFT)); } else { free_bootmem_late(__pa(io_tlb_overflow_buffer), - io_tlb_overflow); + PAGE_ALIGN(io_tlb_overflow)); free_bootmem_late(__pa(io_tlb_orig_addr), - io_tlb_nslabs * sizeof(phys_addr_t)); + PAGE_ALIGN(io_tlb_nslabs * sizeof(phys_addr_t))); free_bootmem_late(__pa(io_tlb_list), - io_tlb_nslabs * sizeof(int)); + PAGE_ALIGN(io_tlb_nslabs * sizeof(int))); free_bootmem_late(__pa(io_tlb_start), - io_tlb_nslabs << IO_TLB_SHIFT); + PAGE_ALIGN(io_tlb_nslabs << IO_TLB_SHIFT)); } } -- 1.6.4.2 From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from rcsinet12.oracle.com ([148.87.113.124]:25872 "EHLO rcsinet12.oracle.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1755231Ab0DIGHM (ORCPT ); Fri, 9 Apr 2010 02:07:12 -0400 From: Yinghai Lu Subject: [PATCH 01/39] swiotlb: Use page alignment for early buffer allocation Date: Thu, 8 Apr 2010 23:03:30 -0700 Message-ID: <1270793048-23796-2-git-send-email-yinghai@kernel.org> In-Reply-To: <1270793048-23796-1-git-send-email-yinghai@kernel.org> References: <1270793048-23796-1-git-send-email-yinghai@kernel.org> Sender: linux-arch-owner@vger.kernel.org List-ID: To: Ingo Molnar , Thomas Gleixner , "H. Peter Anvin" , Andrew Morton Cc: David Miller , Benjamin Herrenschmidt , Linus Torvalds , Johannes Weiner , linux-kernel@vger.kernel.org, linux-arch@vger.kernel.org, Yinghai Lu , FUJITA Tomonori , Becky Bruce Message-ID: <20100409060330.FCkH65lOIvRWkJ9wcaZQw6jlRUq_B1GWnSCxq3j7qng@z> We could call free_bootmem_late() if swiotlb is not used, and it will shrink to page alignement. So alloc them with page alignment at first, to avoid lose two pages before patch: [ 0.000000] lmb_reserve_area: [00d3600000, 00d7600000] swiotlb buffer [ 0.000000] lmb_reserve_area: [00d7e7ef40, 00d7e9ef40] swiotlb list [ 0.000000] lmb_reserve_area: [00d7e3ef40, 00d7e7ef40] swiotlb orig_ad [ 0.000000] lmb_reserve_area: [000008a000, 0000092000] swiotlb overflo after patch will get [ 0.000000] lmb_reserve_area: [00d3600000, 00d7600000] swiotlb buffer [ 0.000000] lmb_reserve_area: [00d7e7e000, 00d7e9e000] swiotlb list [ 0.000000] lmb_reserve_area: [00d7e3e000, 00d7e7e000] swiotlb orig_ad [ 0.000000] lmb_reserve_area: [000008a000, 0000092000] swiotlb overflo Signed-off-by: Yinghai Lu Cc: FUJITA Tomonori Cc: Becky Bruce --- lib/swiotlb.c | 16 ++++++++-------- 1 files changed, 8 insertions(+), 8 deletions(-) diff --git a/lib/swiotlb.c b/lib/swiotlb.c index 5fddf72..1bd4258 100644 --- a/lib/swiotlb.c +++ b/lib/swiotlb.c @@ -159,7 +159,7 @@ swiotlb_init_with_default_size(size_t default_size, int verbose) /* * Get IO TLB memory from the low pages */ - io_tlb_start = alloc_bootmem_low_pages(bytes); + io_tlb_start = alloc_bootmem_low_pages(PAGE_ALIGN(bytes)); if (!io_tlb_start) panic("Cannot allocate SWIOTLB buffer"); io_tlb_end = io_tlb_start + bytes; @@ -169,16 +169,16 @@ swiotlb_init_with_default_size(size_t default_size, int verbose) * to find contiguous free memory regions of size up to IO_TLB_SEGSIZE * between io_tlb_start and io_tlb_end. */ - io_tlb_list = alloc_bootmem(io_tlb_nslabs * sizeof(int)); + io_tlb_list = alloc_bootmem_pages(PAGE_ALIGN(io_tlb_nslabs * sizeof(int))); for (i = 0; i < io_tlb_nslabs; i++) io_tlb_list[i] = IO_TLB_SEGSIZE - OFFSET(i, IO_TLB_SEGSIZE); io_tlb_index = 0; - io_tlb_orig_addr = alloc_bootmem(io_tlb_nslabs * sizeof(phys_addr_t)); + io_tlb_orig_addr = alloc_bootmem_pages(PAGE_ALIGN(io_tlb_nslabs * sizeof(phys_addr_t))); /* * Get the overflow emergency buffer */ - io_tlb_overflow_buffer = alloc_bootmem_low(io_tlb_overflow); + io_tlb_overflow_buffer = alloc_bootmem_low_pages(PAGE_ALIGN(io_tlb_overflow)); if (!io_tlb_overflow_buffer) panic("Cannot allocate SWIOTLB overflow buffer!\n"); if (verbose) @@ -304,13 +304,13 @@ void __init swiotlb_free(void) get_order(io_tlb_nslabs << IO_TLB_SHIFT)); } else { free_bootmem_late(__pa(io_tlb_overflow_buffer), - io_tlb_overflow); + PAGE_ALIGN(io_tlb_overflow)); free_bootmem_late(__pa(io_tlb_orig_addr), - io_tlb_nslabs * sizeof(phys_addr_t)); + PAGE_ALIGN(io_tlb_nslabs * sizeof(phys_addr_t))); free_bootmem_late(__pa(io_tlb_list), - io_tlb_nslabs * sizeof(int)); + PAGE_ALIGN(io_tlb_nslabs * sizeof(int))); free_bootmem_late(__pa(io_tlb_start), - io_tlb_nslabs << IO_TLB_SHIFT); + PAGE_ALIGN(io_tlb_nslabs << IO_TLB_SHIFT)); } } -- 1.6.4.2