From mboxrd@z Thu Jan 1 00:00:00 1970 From: Yinghai Lu Subject: [PATCH 29/37] swiotlb: Use page alignment for early buffer allocation Date: Tue, 15 Jun 2010 22:42:38 -0700 Message-ID: <1276666966-14259-30-git-send-email-yinghai@kernel.org> References: <1276666966-14259-1-git-send-email-yinghai@kernel.org> Return-path: Received: from rcsinet10.oracle.com ([148.87.113.121]:28651 "EHLO rcsinet10.oracle.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1756867Ab0FPFqf (ORCPT ); Wed, 16 Jun 2010 01:46:35 -0400 In-Reply-To: <1276666966-14259-1-git-send-email-yinghai@kernel.org> Sender: linux-arch-owner@vger.kernel.org List-ID: To: Thomas Gleixner , Ingo Molnar , "H. Peter Anvin" , Andrew Morton , Benjamin Herrenschmidt Cc: David Miller , Linus Torvalds , Johannes Weiner , linux-kernel@vger.kernel.org, linux-arch@vger.kernel.org, Yinghai Lu , FUJITA Tomonori , Becky Bruce for 2.6.34 We could call free_bootmem_late() if swiotlb is not used, and it will shrink to page alignement. So alloc them with page alignment at first, to avoid lose two pages before patch: [ 0.000000] lmb_reserve_range: [00d3600000, 00d7600000] swiotlb buffer [ 0.000000] lmb_reserve_range: [00d7e7ef40, 00d7e9ef40] swiotlb list [ 0.000000] lmb_reserve_range: [00d7e3ef40, 00d7e7ef40] swiotlb orig_ad [ 0.000000] lmb_reserve_range: [000008a000, 0000092000] swiotlb overflo after patch will get [ 0.000000] lmb_reserve_range: [00d3600000, 00d7600000] swiotlb buffer [ 0.000000] lmb_reserve_range: [00d7e7e000, 00d7e9e000] swiotlb list [ 0.000000] lmb_reserve_range: [00d7e3e000, 00d7e7e000] swiotlb orig_ad [ 0.000000] lmb_reserve_range: [000008a000, 0000092000] swiotlb overflo Signed-off-by: Yinghai Lu Cc: FUJITA Tomonori Cc: Becky Bruce --- lib/swiotlb.c | 16 ++++++++-------- 1 files changed, 8 insertions(+), 8 deletions(-) diff --git a/lib/swiotlb.c b/lib/swiotlb.c index a009055..ecfab7f 100644 --- a/lib/swiotlb.c +++ b/lib/swiotlb.c @@ -159,7 +159,7 @@ swiotlb_init_with_default_size(size_t default_size, int verbose) /* * Get IO TLB memory from the low pages */ - io_tlb_start = alloc_bootmem_low_pages(bytes); + io_tlb_start = alloc_bootmem_low_pages(PAGE_ALIGN(bytes)); if (!io_tlb_start) panic("Cannot allocate SWIOTLB buffer"); io_tlb_end = io_tlb_start + bytes; @@ -169,16 +169,16 @@ swiotlb_init_with_default_size(size_t default_size, int verbose) * to find contiguous free memory regions of size up to IO_TLB_SEGSIZE * between io_tlb_start and io_tlb_end. */ - io_tlb_list = alloc_bootmem(io_tlb_nslabs * sizeof(int)); + io_tlb_list = alloc_bootmem_pages(PAGE_ALIGN(io_tlb_nslabs * sizeof(int))); for (i = 0; i < io_tlb_nslabs; i++) io_tlb_list[i] = IO_TLB_SEGSIZE - OFFSET(i, IO_TLB_SEGSIZE); io_tlb_index = 0; - io_tlb_orig_addr = alloc_bootmem(io_tlb_nslabs * sizeof(phys_addr_t)); + io_tlb_orig_addr = alloc_bootmem_pages(PAGE_ALIGN(io_tlb_nslabs * sizeof(phys_addr_t))); /* * Get the overflow emergency buffer */ - io_tlb_overflow_buffer = alloc_bootmem_low(io_tlb_overflow); + io_tlb_overflow_buffer = alloc_bootmem_low_pages(PAGE_ALIGN(io_tlb_overflow)); if (!io_tlb_overflow_buffer) panic("Cannot allocate SWIOTLB overflow buffer!\n"); if (verbose) @@ -304,13 +304,13 @@ void __init swiotlb_free(void) get_order(io_tlb_nslabs << IO_TLB_SHIFT)); } else { free_bootmem_late(__pa(io_tlb_overflow_buffer), - io_tlb_overflow); + PAGE_ALIGN(io_tlb_overflow)); free_bootmem_late(__pa(io_tlb_orig_addr), - io_tlb_nslabs * sizeof(phys_addr_t)); + PAGE_ALIGN(io_tlb_nslabs * sizeof(phys_addr_t))); free_bootmem_late(__pa(io_tlb_list), - io_tlb_nslabs * sizeof(int)); + PAGE_ALIGN(io_tlb_nslabs * sizeof(int))); free_bootmem_late(__pa(io_tlb_start), - io_tlb_nslabs << IO_TLB_SHIFT); + PAGE_ALIGN(io_tlb_nslabs << IO_TLB_SHIFT)); } } -- 1.6.4.2 From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from rcsinet10.oracle.com ([148.87.113.121]:28651 "EHLO rcsinet10.oracle.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1756867Ab0FPFqf (ORCPT ); Wed, 16 Jun 2010 01:46:35 -0400 From: Yinghai Lu Subject: [PATCH 29/37] swiotlb: Use page alignment for early buffer allocation Date: Tue, 15 Jun 2010 22:42:38 -0700 Message-ID: <1276666966-14259-30-git-send-email-yinghai@kernel.org> In-Reply-To: <1276666966-14259-1-git-send-email-yinghai@kernel.org> References: <1276666966-14259-1-git-send-email-yinghai@kernel.org> Sender: linux-arch-owner@vger.kernel.org List-ID: To: Thomas Gleixner , Ingo Molnar , "H. Peter Anvin" , Andrew Morton , Benjamin Herrenschmidt Cc: David Miller , Linus Torvalds , Johannes Weiner , linux-kernel@vger.kernel.org, linux-arch@vger.kernel.org, Yinghai Lu , FUJITA Tomonori , Becky Bruce Message-ID: <20100616054238.XTgx8zvHaoVuh6plcl6ksLAP9hZBDGOmbIYct0TwanM@z> for 2.6.34 We could call free_bootmem_late() if swiotlb is not used, and it will shrink to page alignement. So alloc them with page alignment at first, to avoid lose two pages before patch: [ 0.000000] lmb_reserve_range: [00d3600000, 00d7600000] swiotlb buffer [ 0.000000] lmb_reserve_range: [00d7e7ef40, 00d7e9ef40] swiotlb list [ 0.000000] lmb_reserve_range: [00d7e3ef40, 00d7e7ef40] swiotlb orig_ad [ 0.000000] lmb_reserve_range: [000008a000, 0000092000] swiotlb overflo after patch will get [ 0.000000] lmb_reserve_range: [00d3600000, 00d7600000] swiotlb buffer [ 0.000000] lmb_reserve_range: [00d7e7e000, 00d7e9e000] swiotlb list [ 0.000000] lmb_reserve_range: [00d7e3e000, 00d7e7e000] swiotlb orig_ad [ 0.000000] lmb_reserve_range: [000008a000, 0000092000] swiotlb overflo Signed-off-by: Yinghai Lu Cc: FUJITA Tomonori Cc: Becky Bruce --- lib/swiotlb.c | 16 ++++++++-------- 1 files changed, 8 insertions(+), 8 deletions(-) diff --git a/lib/swiotlb.c b/lib/swiotlb.c index a009055..ecfab7f 100644 --- a/lib/swiotlb.c +++ b/lib/swiotlb.c @@ -159,7 +159,7 @@ swiotlb_init_with_default_size(size_t default_size, int verbose) /* * Get IO TLB memory from the low pages */ - io_tlb_start = alloc_bootmem_low_pages(bytes); + io_tlb_start = alloc_bootmem_low_pages(PAGE_ALIGN(bytes)); if (!io_tlb_start) panic("Cannot allocate SWIOTLB buffer"); io_tlb_end = io_tlb_start + bytes; @@ -169,16 +169,16 @@ swiotlb_init_with_default_size(size_t default_size, int verbose) * to find contiguous free memory regions of size up to IO_TLB_SEGSIZE * between io_tlb_start and io_tlb_end. */ - io_tlb_list = alloc_bootmem(io_tlb_nslabs * sizeof(int)); + io_tlb_list = alloc_bootmem_pages(PAGE_ALIGN(io_tlb_nslabs * sizeof(int))); for (i = 0; i < io_tlb_nslabs; i++) io_tlb_list[i] = IO_TLB_SEGSIZE - OFFSET(i, IO_TLB_SEGSIZE); io_tlb_index = 0; - io_tlb_orig_addr = alloc_bootmem(io_tlb_nslabs * sizeof(phys_addr_t)); + io_tlb_orig_addr = alloc_bootmem_pages(PAGE_ALIGN(io_tlb_nslabs * sizeof(phys_addr_t))); /* * Get the overflow emergency buffer */ - io_tlb_overflow_buffer = alloc_bootmem_low(io_tlb_overflow); + io_tlb_overflow_buffer = alloc_bootmem_low_pages(PAGE_ALIGN(io_tlb_overflow)); if (!io_tlb_overflow_buffer) panic("Cannot allocate SWIOTLB overflow buffer!\n"); if (verbose) @@ -304,13 +304,13 @@ void __init swiotlb_free(void) get_order(io_tlb_nslabs << IO_TLB_SHIFT)); } else { free_bootmem_late(__pa(io_tlb_overflow_buffer), - io_tlb_overflow); + PAGE_ALIGN(io_tlb_overflow)); free_bootmem_late(__pa(io_tlb_orig_addr), - io_tlb_nslabs * sizeof(phys_addr_t)); + PAGE_ALIGN(io_tlb_nslabs * sizeof(phys_addr_t))); free_bootmem_late(__pa(io_tlb_list), - io_tlb_nslabs * sizeof(int)); + PAGE_ALIGN(io_tlb_nslabs * sizeof(int))); free_bootmem_late(__pa(io_tlb_start), - io_tlb_nslabs << IO_TLB_SHIFT); + PAGE_ALIGN(io_tlb_nslabs << IO_TLB_SHIFT)); } } -- 1.6.4.2