public inbox for linux-ia64@vger.kernel.org
 help / color / mirror / Atom feed
* [PATCH] long line & codingstyle cleanup for init.c
@ 2005-01-07  0:17 Jesse Barnes
  2005-01-07  0:21 ` [PATCH] long line & codingstyle cleanup for tlb.c Jesse Barnes
  0 siblings, 1 reply; 2+ messages in thread
From: Jesse Barnes @ 2005-01-07  0:17 UTC (permalink / raw)
  To: linux-ia64

[-- Attachment #1: Type: text/plain, Size: 211 bytes --]

Fix long lines and coding style in init.c.  This file could probably use some 
#include trimming as well, but I didn't attempt that with this patch.

Signed-off-by: Jesse Barnes <jbarnes@sgi.com>

Thanks,
Jesse

[-- Attachment #2: init-cleanup.patch --]
[-- Type: text/plain, Size: 18548 bytes --]

===== arch/ia64/mm/init.c 1.74 vs edited =====
--- 1.74/arch/ia64/mm/init.c	2005-01-04 14:58:07 -08:00
+++ edited/arch/ia64/mm/init.c	2005-01-06 16:16:40 -08:00
@@ -7,7 +7,6 @@
 #include <linux/config.h>
 #include <linux/kernel.h>
 #include <linux/init.h>
-
 #include <linux/bootmem.h>
 #include <linux/efi.h>
 #include <linux/elf.h>
@@ -20,7 +19,6 @@
 #include <linux/swap.h>
 #include <linux/proc_fs.h>
 #include <linux/bitops.h>
-
 #include <asm/a.out.h>
 #include <asm/dma.h>
 #include <asm/ia32.h>
@@ -56,8 +54,7 @@
 struct page *zero_page_memmap_ptr;		/* map entry for zero page */
 EXPORT_SYMBOL(zero_page_memmap_ptr);
 
-void
-check_pgt_cache (void)
+void check_pgt_cache(void)
 {
 	int low, high;
 
@@ -76,30 +73,33 @@
 	preempt_enable();
 }
 
-void
-update_mmu_cache (struct vm_area_struct *vma, unsigned long vaddr, pte_t pte)
+void update_mmu_cache(struct vm_area_struct *vma, unsigned long vaddr,
+		      pte_t pte)
 {
 	unsigned long addr;
 	struct page *page;
 
 	if (!pte_exec(pte))
-		return;				/* not an executable page... */
+		return;	/* not an executable page... */
 
 	page = pte_page(pte);
-	/* don't use VADDR: it may not be mapped on this CPU (or may have just been flushed): */
+	/*
+	 * don't use VADDR: it may not be mapped on this CPU (or may have just
+	 * been flushed)
+	 */
 	addr = (unsigned long) page_address(page);
 
 	if (test_bit(PG_arch_1, &page->flags))
-		return;				/* i-cache is already coherent with d-cache */
+		return;	/* i-cache is already coherent with d-cache */
 
 	flush_icache_range(addr, addr + PAGE_SIZE);
-	set_bit(PG_arch_1, &page->flags);	/* mark page as clean */
+	set_bit(PG_arch_1, &page->flags); /* mark page as clean */
 }
 
-inline void
-ia64_set_rbs_bot (void)
+inline void ia64_set_rbs_bot(void)
 {
-	unsigned long stack_size = current->signal->rlim[RLIMIT_STACK].rlim_max & -16;
+	unsigned long stack_size =
+		current->signal->rlim[RLIMIT_STACK].rlim_max & -16;
 
 	if (stack_size > MAX_USER_STACK_SIZE)
 		stack_size = MAX_USER_STACK_SIZE;
@@ -112,17 +112,17 @@
  * store (which grows upwards) and install the gateway page which is
  * used for signal trampolines, etc.
  */
-void
-ia64_init_addr_space (void)
+void ia64_init_addr_space(void)
 {
 	struct vm_area_struct *vma;
 
 	ia64_set_rbs_bot();
 
 	/*
-	 * If we're out of memory and kmem_cache_alloc() returns NULL, we simply ignore
-	 * the problem.  When the process attempts to write to the register backing store
-	 * for the first time, it will get a SEGFAULT in this case.
+	 * If we're out of memory and kmem_cache_alloc() returns NULL, we
+	 * simply ignore the problem.  When the process attempts to write to
+	 * the register backing store for the first time, it will get a
+	 * SEGFAULT in this case.
 	 */
 	vma = kmem_cache_alloc(vm_area_cachep, SLAB_KERNEL);
 	if (vma) {
@@ -130,7 +130,8 @@
 		vma->vm_mm = current->mm;
 		vma->vm_start = current->thread.rbs_bot & PAGE_MASK;
 		vma->vm_end = vma->vm_start + PAGE_SIZE;
-		vma->vm_page_prot = protection_map[VM_DATA_DEFAULT_FLAGS & 0x7];
+		vma->vm_page_prot =
+			protection_map[VM_DATA_DEFAULT_FLAGS & 0x7];
 		vma->vm_flags = VM_DATA_DEFAULT_FLAGS | VM_GROWSUP;
 		down_write(&current->mm->mmap_sem);
 		if (insert_vm_struct(current->mm, vma)) {
@@ -141,15 +142,21 @@
 		up_write(&current->mm->mmap_sem);
 	}
 
-	/* map NaT-page at address zero to speed up speculative dereferencing of NULL: */
+	/*
+	 * map NaT-page at address zero to speed up speculative dereferencing
+	 * of NULL
+	 */
 	if (!(current->personality & MMAP_PAGE_ZERO)) {
 		vma = kmem_cache_alloc(vm_area_cachep, SLAB_KERNEL);
 		if (vma) {
 			memset(vma, 0, sizeof(*vma));
 			vma->vm_mm = current->mm;
 			vma->vm_end = PAGE_SIZE;
-			vma->vm_page_prot = __pgprot(pgprot_val(PAGE_READONLY) | _PAGE_MA_NAT);
-			vma->vm_flags = VM_READ | VM_MAYREAD | VM_IO | VM_RESERVED;
+			vma->vm_page_prot =
+				__pgprot(pgprot_val(PAGE_READONLY) |
+					 _PAGE_MA_NAT);
+			vma->vm_flags = VM_READ | VM_MAYREAD | VM_IO |
+				VM_RESERVED;
 			down_write(&current->mm->mmap_sem);
 			if (insert_vm_struct(current->mm, vma)) {
 				up_write(&current->mm->mmap_sem);
@@ -161,8 +168,7 @@
 	}
 }
 
-void
-free_initmem (void)
+void free_initmem(void)
 {
 	unsigned long addr, eaddr;
 
@@ -179,8 +185,7 @@
 	       (__init_end - __init_begin) >> 10);
 }
 
-void
-free_initrd_mem (unsigned long start, unsigned long end)
+void free_initrd_mem(unsigned long start, unsigned long end)
 {
 	struct page *page;
 	/*
@@ -218,7 +223,8 @@
 	end = end & PAGE_MASK;
 
 	if (start < end)
-		printk(KERN_INFO "Freeing initrd memory: %ldkB freed\n", (end - start) >> 10);
+		printk(KERN_INFO "Freeing initrd memory: %ldkB freed\n",
+		       (end - start) >> 10);
 
 	for (; start < end; start += PAGE_SIZE) {
 		if (!virt_addr_valid(start))
@@ -234,8 +240,8 @@
 /*
  * This installs a clean page in the kernel's page table.
  */
-struct page *
-put_kernel_page (struct page *page, unsigned long address, pgprot_t pgprot)
+struct page *put_kernel_page(struct page *page, unsigned long address,
+			     pgprot_t pgprot)
 {
 	pgd_t *pgd;
 	pud_t *pud;
@@ -243,10 +249,10 @@
 	pte_t *pte;
 
 	if (!PageReserved(page))
-		printk(KERN_ERR "put_kernel_page: page at 0x%p not in reserved memory\n",
-		       page_address(page));
+		printk(KERN_ERR "put_kernel_page: page at 0x%p not in "
+		       "reserved memory\n", page_address(page));
 
-	pgd = pgd_offset_k(address);		/* note: this is NOT pgd_offset()! */
+	pgd = pgd_offset_k(address); /* note: this is NOT pgd_offset()! */
 
 	spin_lock(&init_mm.page_table_lock);
 	{
@@ -272,14 +278,14 @@
 	return page;
 }
 
-static void
-setup_gate (void)
+static void setup_gate(void)
 {
 	struct page *page;
 
 	/*
-	 * Map the gate page twice: once read-only to export the ELF headers etc. and once
-	 * execute-only page to enable privilege-promotion via "epc":
+	 * Map the gate page twice: once read-only to export the ELF headers
+	 * etc. and once execute-only page to enable privilege-promotion via
+	 * "epc":
 	 */
 	page = virt_to_page(ia64_imva(__start_gate_section));
 	put_kernel_page(page, GATE_ADDR, PAGE_READONLY);
@@ -292,8 +298,7 @@
 	ia64_patch_gate();
 }
 
-void __devinit
-ia64_mmu_init (void *my_cpu_data)
+void __devinit ia64_mmu_init(void *my_cpu_data)
 {
 	unsigned long psr, pta, impl_va_bits;
 	extern void __devinit tlb_init (void);
@@ -315,39 +320,43 @@
 	ia64_srlz_i();
 
 	/*
-	 * Check if the virtually mapped linear page table (VMLPT) overlaps with a mapped
-	 * address space.  The IA-64 architecture guarantees that at least 50 bits of
-	 * virtual address space are implemented but if we pick a large enough page size
-	 * (e.g., 64KB), the mapped address space is big enough that it will overlap with
-	 * VMLPT.  I assume that once we run on machines big enough to warrant 64KB pages,
-	 * IMPL_VA_MSB will be significantly bigger, so this is unlikely to become a
-	 * problem in practice.  Alternatively, we could truncate the top of the mapped
-	 * address space to not permit mappings that would overlap with the VMLPT.
+	 * Check if the virtually mapped linear page table (VMLPT) overlaps
+	 * with a mapped address space.  The IA-64 architecture guarantees that
+	 * at least 50 bits of virtual address space are implemented but if we
+	 * pick a large enough page size (e.g., 64KB), the mapped address space
+	 * is big enough that it will overlap with VMLPT.  I assume that once
+	 * we run on machines big enough to warrant 64KB pages, IMPL_VA_MSB
+	 * will be significantly bigger, so this is unlikely to become a
+	 * problem in practice.  Alternatively, we could truncate the top of
+	 * the mapped address space to not permit mappings that would overlap
+	 * with the VMLPT.
 	 * --davidm 00/12/06
 	 */
-#	define pte_bits			3
-#	define mapped_space_bits	(3*(PAGE_SHIFT - pte_bits) + PAGE_SHIFT)
+#	define pte_bits          3
+#	define mapped_space_bits (3*(PAGE_SHIFT - pte_bits) + PAGE_SHIFT)
 	/*
-	 * The virtual page table has to cover the entire implemented address space within
-	 * a region even though not all of this space may be mappable.  The reason for
-	 * this is that the Access bit and Dirty bit fault handlers perform
-	 * non-speculative accesses to the virtual page table, so the address range of the
-	 * virtual page table itself needs to be covered by virtual page table.
+	 * The virtual page table has to cover the entire implemented address
+	 * space within a region even though not all of this space may be
+	 * mappable.  The reason for this is that the Access bit and Dirty bit
+	 * fault handlers perform non-speculative accesses to the virtual page
+	 * table, so the address range of the virtual page table itself needs
+	 * to be covered by virtual page table.
 	 */
-#	define vmlpt_bits		(impl_va_bits - PAGE_SHIFT + pte_bits)
-#	define POW2(n)			(1ULL << (n))
+#	define vmlpt_bits (impl_va_bits - PAGE_SHIFT + pte_bits)
+#	define POW2(n)    (1ULL << (n))
 
 	impl_va_bits = ffz(~(local_cpu_data->unimpl_va_mask | (7UL << 61)));
 
 	if (impl_va_bits < 51 || impl_va_bits > 61)
-		panic("CPU has bogus IMPL_VA_MSB value of %lu!\n", impl_va_bits - 1);
+		panic("CPU has bogus IMPL_VA_MSB value of %lu!\n",
+		      impl_va_bits - 1);
 
 	/* place the VMLPT at the end of each page-table mapped region: */
 	pta = POW2(61) - POW2(vmlpt_bits);
 
 	if (POW2(mapped_space_bits) >= pta)
-		panic("mm/init: overlap between virtually mapped linear page table and "
-		      "mapped kernel space!");
+		panic("mm/init: overlap between virtually mapped linear page "
+		      "table and mapped kernel space!");
 	/*
 	 * Set the (virtually mapped linear) page table address.  Bit
 	 * 8 selects between the short and long format, bits 2-7 the
@@ -364,14 +373,15 @@
 #endif
 
 	/*
-	 * The MCA info structure was allocated earlier and a physical address pointer
-	 * saved in k3.  Move that pointer into the cpuinfo structure and save
-	 * the physical address of the cpuinfo structure in k3.
+	 * The MCA info structure was allocated earlier and a physical address
+	 * pointer saved in k3.  Move that pointer into the cpuinfo structure
+	 * and save the physical address of the cpuinfo structure in k3.
 	 */
 	cpuinfo = (struct cpuinfo_ia64 *)my_cpu_data;
 	cpuinfo->ia64_pa_mca_data = (__u64 *)ia64_get_kr(IA64_KR_PA_CPU_INFO);
 
-	cpuinfo->percpu_paddr = pte_val(mk_pte_phys(__pa(my_cpu_data), PAGE_KERNEL));
+	cpuinfo->percpu_paddr = pte_val(mk_pte_phys(__pa(my_cpu_data),
+						    PAGE_KERNEL));
 	ia64_set_kr(IA64_KR_PA_CPU_INFO, __pa(my_cpu_data));
 
 	/*
@@ -381,9 +391,7 @@
 }
 
 #ifdef CONFIG_VIRTUAL_MEM_MAP
-
-int
-create_mem_map_page_table (u64 start, u64 end, void *arg)
+int create_mem_map_page_table(u64 start, u64 end, void *arg)
 {
 	unsigned long address, start_page, end_page;
 	struct page *map_start, *map_end;
@@ -403,20 +411,24 @@
 	for (address = start_page; address < end_page; address += PAGE_SIZE) {
 		pgd = pgd_offset_k(address);
 		if (pgd_none(*pgd))
-			pgd_populate(&init_mm, pgd, alloc_bootmem_pages_node(NODE_DATA(node), PAGE_SIZE));
+			pgd_populate(&init_mm, pgd,
+				     alloc_bootmem_pages_node(NODE_DATA(node),
+							      PAGE_SIZE));
 		pud = pud_offset(pgd, address);
 
 		if (pud_none(*pud))
-			pud_populate(&init_mm, pud, alloc_bootmem_pages_node(NODE_DATA(node), PAGE_SIZE));
+			pud_populate(&init_mm, pud,
+				     alloc_bootmem_pages_node(NODE_DATA(node),
+							      PAGE_SIZE));
 		pmd = pmd_offset(pud, address);
 
 		if (pmd_none(*pmd))
 			pmd_populate_kernel(&init_mm, pmd, alloc_bootmem_pages_node(NODE_DATA(node), PAGE_SIZE));
+
 		pte = pte_offset_kernel(pmd, address);
 
 		if (pte_none(*pte))
-			set_pte(pte, pfn_pte(__pa(alloc_bootmem_pages_node(NODE_DATA(node), PAGE_SIZE)) >> PAGE_SHIFT,
-					     PAGE_KERNEL));
+			set_pte(pte, pfn_pte(__pa(alloc_bootmem_pages_node(NODE_DATA(node), PAGE_SIZE)) >> PAGE_SHIFT, PAGE_KERNEL));
 	}
 	return 0;
 }
@@ -428,8 +440,7 @@
 	unsigned long zone;
 };
 
-static int
-virtual_memmap_init (u64 start, u64 end, void *arg)
+static int virtual_memmap_init(u64 start, u64 end, void *arg)
 {
 	struct memmap_init_callback_data *args;
 	struct page *map_start, *map_end;
@@ -445,23 +456,25 @@
 		map_end = args->end;
 
 	/*
-	 * We have to initialize "out of bounds" struct page elements that fit completely
-	 * on the same pages that were allocated for the "in bounds" elements because they
-	 * may be referenced later (and found to be "reserved").
-	 */
-	map_start -= ((unsigned long) map_start & (PAGE_SIZE - 1)) / sizeof(struct page);
-	map_end += ((PAGE_ALIGN((unsigned long) map_end) - (unsigned long) map_end)
-		    / sizeof(struct page));
+	 * We have to initialize "out of bounds" struct page elements that fit
+	 * completely on the same pages that were allocated for the "in bounds"
+	 * elements because they may be referenced later (and found to be
+	 * "reserved").
+	 */
+	map_start -= ((unsigned long) map_start & (PAGE_SIZE - 1)) /
+		sizeof(struct page);
+	map_end += (PAGE_ALIGN((unsigned long) map_end) -
+		    (unsigned long) map_end) / sizeof(struct page);
 
 	if (map_start < map_end)
 		memmap_init_zone((unsigned long)(map_end - map_start),
-				 args->nid, args->zone, page_to_pfn(map_start));
+				 args->nid, args->zone,
+				 page_to_pfn(map_start));
 	return 0;
 }
 
-void
-memmap_init (unsigned long size, int nid, unsigned long zone,
-	     unsigned long start_pfn)
+void memmap_init(unsigned long size, int nid, unsigned long zone,
+		 unsigned long start_pfn)
 {
 	if (!vmem_map)
 		memmap_init_zone(size, nid, zone, start_pfn);
@@ -479,20 +492,18 @@
 	}
 }
 
-int
-ia64_pfn_valid (unsigned long pfn)
+int ia64_pfn_valid(unsigned long pfn)
 {
-	char byte;
 	struct page *pg = pfn_to_page(pfn);
+	char byte;
 
-	return     (__get_user(byte, (char __user *) pg) == 0)
-		&& ((((u64)pg & PAGE_MASK) == (((u64)(pg + 1) - 1) & PAGE_MASK))
-			|| (__get_user(byte, (char __user *) (pg + 1) - 1) == 0));
+	return (__get_user(byte, (char __user *) pg) == 0) &&
+		((((u64)pg & PAGE_MASK) == (((u64)(pg + 1) - 1) & PAGE_MASK))
+		 || (__get_user(byte, (char __user *) (pg + 1) - 1) == 0));
 }
 EXPORT_SYMBOL(ia64_pfn_valid);
 
-int
-find_largest_hole (u64 start, u64 end, void *arg)
+int find_largest_hole(u64 start, u64 end, void *arg)
 {
 	u64 *max_gap = arg;
 
@@ -507,8 +518,7 @@
 }
 #endif /* CONFIG_VIRTUAL_MEM_MAP */
 
-static int
-count_reserved_pages (u64 start, u64 end, void *arg)
+static int count_reserved_pages(u64 start, u64 end, void *arg)
 {
 	unsigned long num_reserved = 0;
 	unsigned long *count = arg;
@@ -521,17 +531,15 @@
 }
 
 /*
- * Boot command-line option "nolwsys" can be used to disable the use of any light-weight
- * system call handler.  When this option is in effect, all fsyscalls will end up bubbling
- * down into the kernel and calling the normal (heavy-weight) syscall handler.  This is
- * useful for performance testing, but conceivably could also come in handy for debugging
- * purposes.
+ * Boot command-line option "nolwsys" can be used to disable the use of any
+ * light-weight system call handler.  When this option is in effect, all
+ * fsyscalls will end up bubbling down into the kernel and calling the normal
+ * (heavy-weight) syscall handler.  This is useful for performance testing, but
+ * conceivably could also come in handy for debugging purposes.
  */
-
 static int nolwsys;
 
-static int __init
-nolwsys_setup (char *s)
+static int __init nolwsys_setup(char *s)
 {
 	nolwsys = 1;
 	return 1;
@@ -539,8 +547,7 @@
 
 __setup("nolwsys", nolwsys_setup);
 
-void
-mem_init (void)
+void mem_init(void)
 {
 	long reserved_pages, codesize, datasize, initsize;
 	unsigned long num_pgt_pages;
@@ -550,9 +557,9 @@
 
 #ifdef CONFIG_PCI
 	/*
-	 * This needs to be called _after_ the command line has been parsed but _before_
-	 * any drivers that may need the PCI DMA interface are initialized or bootmem has
-	 * been freed.
+	 * This needs to be called _after_ the command line has been parsed but
+	 * _before_ any drivers that may need the PCI DMA interface are
+	 * initialized or bootmem has been freed.
 	 */
 	platform_dma_init();
 #endif
@@ -566,7 +573,8 @@
 	high_memory = __va(max_low_pfn * PAGE_SIZE);
 
 	kclist_add(&kcore_mem, __va(0), max_low_pfn * PAGE_SIZE);
-	kclist_add(&kcore_vmem, (void *)VMALLOC_START, VMALLOC_END-VMALLOC_START);
+	kclist_add(&kcore_vmem, (void *)VMALLOC_START,
+		   VMALLOC_END - VMALLOC_START);
 	kclist_add(&kcore_kernel, _stext, _end - _stext);
 
 	for_each_pgdat(pgdat)
@@ -579,16 +587,24 @@
 	datasize =  (unsigned long) _edata - (unsigned long) _etext;
 	initsize =  (unsigned long) __init_end - (unsigned long) __init_begin;
 
-	printk(KERN_INFO "Memory: %luk/%luk available (%luk code, %luk reserved, "
-	       "%luk data, %luk init)\n", (unsigned long) nr_free_pages() << (PAGE_SHIFT - 10),
-	       num_physpages << (PAGE_SHIFT - 10), codesize >> 10,
-	       reserved_pages << (PAGE_SHIFT - 10), datasize >> 10, initsize >> 10);
-
-	/*
-	 * Allow for enough (cached) page table pages so that we can map the entire memory
-	 * at least once.  Each task also needs a couple of page tables pages, so add in a
-	 * fudge factor for that (don't use "threads-max" here; that would be wrong!).
-	 * Don't allow the cache to be more than 10% of total memory, though.
+	printk(KERN_INFO "Memory: %luk/%luk available "
+	       "(%luk code, "
+	       "%luk reserved, "
+	       "%luk data, "
+	       "%luk init)\n",
+	       (unsigned long) nr_free_pages() << (PAGE_SHIFT - 10),
+	       num_physpages << (PAGE_SHIFT - 10),
+	       codesize >> 10,
+	       reserved_pages << (PAGE_SHIFT - 10),
+	       datasize >> 10,
+	       initsize >> 10);
+
+	/*
+	 * Allow for enough (cached) page table pages so that we can map the
+	 * entire memory at least once.  Each task also needs a couple of page
+	 * tables pages, so add in a fudge factor for that (don't use
+	 * "threads-max" here; that would be wrong!). Don't allow the cache to
+	 * be more than 10% of total memory, though.
 	 */
 #	define NUM_TASKS	500	/* typical number of tasks */
 	num_pgt_pages = nr_free_pages() / PTRS_PER_PGD + NUM_TASKS;
@@ -598,8 +614,9 @@
 		pgt_cache_water[1] = num_pgt_pages;
 
 	/*
-	 * For fsyscall entrpoints with no light-weight handler, use the ordinary
-	 * (heavy-weight) handler, but mark it by setting bit 0, so the fsyscall entry
+	 * For fsyscall entrpoints with no light-weight handler, use the
+	 * ordinary (heavy-weight) handler, but mark it by setting bit 0, so
+	 * the fsyscall entry
 	 * code can tell them apart.
 	 */
 	for (i = 0; i < NR_syscalls; ++i) {

^ permalink raw reply	[flat|nested] 2+ messages in thread

* [PATCH] long line & codingstyle cleanup for tlb.c
  2005-01-07  0:17 [PATCH] long line & codingstyle cleanup for init.c Jesse Barnes
@ 2005-01-07  0:21 ` Jesse Barnes
  0 siblings, 0 replies; 2+ messages in thread
From: Jesse Barnes @ 2005-01-07  0:21 UTC (permalink / raw)
  To: linux-ia64

[-- Attachment #1: Type: text/plain, Size: 106 bytes --]

Fixup long lines and coding style in tlb.c.

Signed-off-by: Jesse Barnes <jbarnes@sgi.com>

Thanks,
Jesse

[-- Attachment #2: tlb-cleanup.patch --]
[-- Type: text/plain, Size: 5047 bytes --]

===== arch/ia64/mm/tlb.c 1.23 vs edited =====
--- 1.23/arch/ia64/mm/tlb.c	2004-10-05 11:24:11 -07:00
+++ edited/arch/ia64/mm/tlb.c	2005-01-06 16:11:49 -08:00
@@ -16,7 +16,6 @@
 #include <linux/sched.h>
 #include <linux/smp.h>
 #include <linux/mm.h>
-
 #include <asm/delay.h>
 #include <asm/mmu_context.h>
 #include <asm/pgalloc.h>
@@ -25,13 +24,15 @@
 
 static struct {
 	unsigned long mask;	/* mask of supported purge page-sizes */
-	unsigned long max_bits;	/* log2() of largest supported purge page-size */
+	unsigned long max_bits;	/* log2() of largest supported purge
+				   page-size */
 } purge;
 
 struct ia64_ctx ia64_ctx = {
 	.lock =		SPIN_LOCK_UNLOCKED,
 	.next =		1,
-	.limit =	(1 << 15) - 1,		/* start out with the safe (architected) limit */
+	.limit =	(1 << 15) - 1, /* start out with the safe
+					  (architected) limit */
 	.max_ctx =	~0U
 };
 
@@ -40,12 +41,11 @@
 /*
  * Acquire the ia64_ctx.lock before calling this function!
  */
-void
-wrap_mmu_context (struct mm_struct *mm)
+void wrap_mmu_context(struct mm_struct *mm)
 {
 	unsigned long tsk_context, max_ctx = ia64_ctx.max_ctx;
 	struct task_struct *tsk;
-	int i;
+	int i, cpu;
 
 	if (ia64_ctx.next > max_ctx)
 		ia64_ctx.next = 300;	/* skip daemons */
@@ -54,7 +54,6 @@
 	/*
 	 * Scan all the task's mm->context and set proper safe range
 	 */
-
 	read_lock(&tasklist_lock);
   repeat:
 	for_each_process(tsk) {
@@ -63,30 +62,33 @@
 		tsk_context = tsk->mm->context;
 		if (tsk_context == ia64_ctx.next) {
 			if (++ia64_ctx.next >= ia64_ctx.limit) {
-				/* empty range: reset the range limit and start over */
+				/* empty range: reset the range limit and
+				   start over */
 				if (ia64_ctx.next > max_ctx)
 					ia64_ctx.next = 300;
 				ia64_ctx.limit = max_ctx + 1;
 				goto repeat;
 			}
 		}
-		if ((tsk_context > ia64_ctx.next) && (tsk_context < ia64_ctx.limit))
+		if ((tsk_context > ia64_ctx.next) &&
+		    (tsk_context < ia64_ctx.limit))
 			ia64_ctx.limit = tsk_context;
 	}
 	read_unlock(&tasklist_lock);
-	/* can't call flush_tlb_all() here because of race condition with O(1) scheduler [EF] */
-	{
-		int cpu = get_cpu(); /* prevent preemption/migration */
-		for (i = 0; i < NR_CPUS; ++i)
-			if (cpu_online(i) && (i != cpu))
-				per_cpu(ia64_need_tlb_flush, i) = 1;
-		put_cpu();
-	}
+	/*
+	 * can't call flush_tlb_all() here because of race condition with
+	 * O(1) scheduler [EF]
+	 */
+	cpu = get_cpu(); /* prevent preemption/migration */
+	for (i = 0; i < NR_CPUS; ++i)
+		if (cpu_online(i) && (i != cpu))
+			per_cpu(ia64_need_tlb_flush, i) = 1;
+	put_cpu();
 	local_flush_tlb_all();
 }
 
-void
-ia64_global_tlb_purge (unsigned long start, unsigned long end, unsigned long nbits)
+void ia64_global_tlb_purge(unsigned long start, unsigned long end,
+			   unsigned long nbits)
 {
 	static spinlock_t ptcg_lock = SPIN_LOCK_UNLOCKED;
 
@@ -105,8 +107,7 @@
 	spin_unlock(&ptcg_lock);
 }
 
-void
-local_flush_tlb_all (void)
+void local_flush_tlb_all(void)
 {
 	unsigned long i, j, flags, count0, count1, stride0, stride1, addr;
 
@@ -129,15 +130,16 @@
 }
 EXPORT_SYMBOL(local_flush_tlb_all);
 
-void
-flush_tlb_range (struct vm_area_struct *vma, unsigned long start, unsigned long end)
+void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
+		     unsigned long end)
 {
 	struct mm_struct *mm = vma->vm_mm;
 	unsigned long size = end - start;
 	unsigned long nbits;
 
 	if (mm != current->active_mm) {
-		/* this does happen, but perhaps it's not worth optimizing for? */
+		/* this does happen, but perhaps it's not worth optimizing
+		   for? */
 #ifdef CONFIG_SMP
 		flush_tlb_all();
 #else
@@ -147,27 +149,27 @@
 	}
 
 	nbits = ia64_fls(size + 0xfff);
-	while (unlikely (((1UL << nbits) & purge.mask) == 0) && (nbits < purge.max_bits))
+	while (unlikely (((1UL << nbits) & purge.mask) == 0) &&
+	       (nbits < purge.max_bits))
 		++nbits;
 	if (nbits > purge.max_bits)
 		nbits = purge.max_bits;
 	start &= ~((1UL << nbits) - 1);
 
-# ifdef CONFIG_SMP
+#ifdef CONFIG_SMP
 	platform_global_tlb_purge(start, end, nbits);
-# else
+#else
 	do {
 		ia64_ptcl(start, (nbits<<2));
 		start += (1UL << nbits);
 	} while (start < end);
-# endif
+#endif
 
 	ia64_srlz_i();			/* srlz.i implies srlz.d */
 }
 EXPORT_SYMBOL(flush_tlb_range);
 
-void __devinit
-ia64_tlb_init (void)
+void __devinit ia64_tlb_init(void)
 {
 	ia64_ptce_info_t ptce_info;
 	unsigned long tr_pgbits;
@@ -175,7 +177,8 @@
 
 	if ((status = ia64_pal_vm_page_size(&tr_pgbits, &purge.mask)) != 0) {
 		printk(KERN_ERR "PAL_VM_PAGE_SIZE failed with status=%ld;"
-		       "defaulting to architected purge page-sizes.\n", status);
+		       "defaulting to architected purge page-sizes.\n",
+		       status);
 		purge.mask = 0x115557000UL;
 	}
 	purge.max_bits = ia64_fls(purge.mask);
@@ -187,5 +190,5 @@
 	local_cpu_data->ptce_stride[0] = ptce_info.stride[0];
 	local_cpu_data->ptce_stride[1] = ptce_info.stride[1];
 
-	local_flush_tlb_all();		/* nuke left overs from bootstrapping... */
+	local_flush_tlb_all(); /* nuke left overs from bootstrapping... */
 }

^ permalink raw reply	[flat|nested] 2+ messages in thread

end of thread, other threads:[~2005-01-07  0:21 UTC | newest]

Thread overview: 2+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2005-01-07  0:17 [PATCH] long line & codingstyle cleanup for init.c Jesse Barnes
2005-01-07  0:21 ` [PATCH] long line & codingstyle cleanup for tlb.c Jesse Barnes

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox