From: Julien Grall <julien.grall@arm.com>
To: xen-devel@lists.xen.org
Cc: George Dunlap <george.dunlap@eu.citrix.com>,
Andrew Cooper <andrew.cooper3@citrix.com>,
Julien Grall <julien.grall@arm.com>,
Jan Beulich <jbeulich@suse.com>
Subject: [PATCH 02/15] xen/x86: p2m-pod: Remove trailing whitespaces
Date: Wed, 13 Sep 2017 18:59:40 +0100 [thread overview]
Message-ID: <20170913175953.16942-3-julien.grall@arm.com> (raw)
In-Reply-To: <20170913175953.16942-1-julien.grall@arm.com>
Signed-off-by: Julien Grall <julien.grall@arm.com>
---
Cc: George Dunlap <george.dunlap@eu.citrix.com>
Cc: Jan Beulich <jbeulich@suse.com>
Cc: Andrew Cooper <andrew.cooper3@citrix.com>
---
xen/arch/x86/mm/p2m-pod.c | 46 +++++++++++++++++++++++-----------------------
1 file changed, 23 insertions(+), 23 deletions(-)
diff --git a/xen/arch/x86/mm/p2m-pod.c b/xen/arch/x86/mm/p2m-pod.c
index fec87e5224..1f07441259 100644
--- a/xen/arch/x86/mm/p2m-pod.c
+++ b/xen/arch/x86/mm/p2m-pod.c
@@ -1,7 +1,7 @@
/******************************************************************************
* arch/x86/mm/p2m-pod.c
*
- * Populate-on-demand p2m entries.
+ * Populate-on-demand p2m entries.
*
* Copyright (c) 2009-2011 Citrix Systems, Inc.
*
@@ -76,7 +76,7 @@ p2m_pod_cache_add(struct p2m_domain *p2m,
__func__, mfn_x(mfn), order, ((1UL << order) - 1));
return -1;
}
-
+
for(i=0; i < 1 << order ; i++) {
struct domain * od;
@@ -223,8 +223,8 @@ p2m_pod_set_cache_target(struct p2m_domain *p2m, unsigned long pod_target, int p
/* If we can't allocate a superpage, try singleton pages */
order = PAGE_ORDER_4K;
goto retry;
- }
-
+ }
+
printk("%s: Unable to allocate page for PoD cache (target=%lu cache=%ld)\n",
__func__, pod_target, p2m->pod.count);
ret = -ENOMEM;
@@ -272,7 +272,7 @@ p2m_pod_set_cache_target(struct p2m_domain *p2m, unsigned long pod_target, int p
if ( test_and_clear_bit(_PGT_pinned, &(page+i)->u.inuse.type_info) )
put_page_and_type(page+i);
-
+
if ( test_and_clear_bit(_PGC_allocated, &(page+i)->count_info) )
put_page(page+i);
@@ -296,7 +296,7 @@ out:
* definitions:
* + M: static_max
* + B: number of pages the balloon driver has ballooned down to.
- * + P: Number of populated pages.
+ * + P: Number of populated pages.
* + T: Old target
* + T': New target
*
@@ -311,10 +311,10 @@ out:
* the remainder of the ram to the guest OS.
* T <T'<B : Increase PoD cache size.
* T'<T<=B : Here we have a choice. We can decrease the size of the cache,
- * get the memory right away. However, that means every time we
- * reduce the memory target we risk the guest attempting to populate the
+ * get the memory right away. However, that means every time we
+ * reduce the memory target we risk the guest attempting to populate the
* memory before the balloon driver has reached its new target. Safer to
- * never reduce the cache size here, but only when the balloon driver frees
+ * never reduce the cache size here, but only when the balloon driver frees
* PoD ranges.
*
* If there are many zero pages, we could reach the target also by doing
@@ -511,7 +511,7 @@ p2m_pod_decrease_reservation(struct domain *d,
long pod, nonpod, ram;
gfn_lock(p2m, gpfn, order);
- pod_lock(p2m);
+ pod_lock(p2m);
/* If we don't have any outstanding PoD entries, let things take their
* course */
@@ -629,7 +629,7 @@ p2m_pod_decrease_reservation(struct domain *d,
nonpod -= n;
ram -= n;
}
- }
+ }
/* If there are no more non-PoD entries, tell decrease_reservation() that
* there's nothing left to do. */
@@ -682,7 +682,7 @@ p2m_pod_zero_check_superpage(struct p2m_domain *p2m, unsigned long gfn)
if ( paging_mode_shadow(d) )
max_ref++;
- /* NOTE: this is why we don't enforce deadlock constraints between p2m
+ /* NOTE: this is why we don't enforce deadlock constraints between p2m
* and pod locks */
gfn_lock(p2m, gfn, SUPERPAGE_ORDER);
@@ -690,7 +690,7 @@ p2m_pod_zero_check_superpage(struct p2m_domain *p2m, unsigned long gfn)
* and aligned, and mapping them. */
for ( i = 0; i < SUPERPAGE_PAGES; i += n )
{
- p2m_access_t a;
+ p2m_access_t a;
unsigned int cur_order;
unsigned long k;
const struct page_info *page;
@@ -807,7 +807,7 @@ p2m_pod_zero_check_superpage(struct p2m_domain *p2m, unsigned long gfn)
out_reset:
if ( reset )
p2m_set_entry(p2m, gfn, mfn0, 9, type0, p2m->default_access);
-
+
out:
gfn_unlock(p2m, gfn, SUPERPAGE_ORDER);
return ret;
@@ -836,8 +836,8 @@ p2m_pod_zero_check(struct p2m_domain *p2m, unsigned long *gfns, int count)
/* If this is ram, and not a pagetable or from the xen heap, and probably not mapped
elsewhere, map it; otherwise, skip. */
if ( p2m_is_ram(types[i])
- && ( (mfn_to_page(mfns[i])->count_info & PGC_allocated) != 0 )
- && ( (mfn_to_page(mfns[i])->count_info & (PGC_page_table|PGC_xen_heap)) == 0 )
+ && ( (mfn_to_page(mfns[i])->count_info & PGC_allocated) != 0 )
+ && ( (mfn_to_page(mfns[i])->count_info & (PGC_page_table|PGC_xen_heap)) == 0 )
&& ( (mfn_to_page(mfns[i])->count_info & PGC_count_mask) <= max_ref ) )
map[i] = map_domain_page(mfns[i]);
else
@@ -915,7 +915,7 @@ p2m_pod_zero_check(struct p2m_domain *p2m, unsigned long *gfns, int count)
t.mfn = mfn_x(mfns[i]);
t.d = d->domain_id;
t.order = 0;
-
+
__trace_var(TRC_MEM_POD_ZERO_RECLAIM, 0, sizeof(t), &t);
}
@@ -924,7 +924,7 @@ p2m_pod_zero_check(struct p2m_domain *p2m, unsigned long *gfns, int count)
p2m->pod.entry_count++;
}
}
-
+
}
#define POD_SWEEP_LIMIT 1024
@@ -1046,12 +1046,12 @@ p2m_pod_demand_populate(struct p2m_domain *p2m, unsigned long gfn,
pod_lock(p2m);
/* This check is done with the pod lock held. This will make sure that
- * even if d->is_dying changes under our feet, p2m_pod_empty_cache()
+ * even if d->is_dying changes under our feet, p2m_pod_empty_cache()
* won't start until we're done. */
if ( unlikely(d->is_dying) )
goto out_fail;
-
+
/* Because PoD does not have cache list for 1GB pages, it has to remap
* 1GB region to 2MB chunks for a retry. */
if ( order == PAGE_ORDER_1G )
@@ -1107,7 +1107,7 @@ p2m_pod_demand_populate(struct p2m_domain *p2m, unsigned long gfn,
set_gpfn_from_mfn(mfn_x(mfn) + i, gfn_aligned + i);
paging_mark_dirty(d, mfn_add(mfn, i));
}
-
+
p2m->pod.entry_count -= (1 << order);
BUG_ON(p2m->pod.entry_count < 0);
@@ -1124,7 +1124,7 @@ p2m_pod_demand_populate(struct p2m_domain *p2m, unsigned long gfn,
t.mfn = mfn_x(mfn);
t.d = d->domain_id;
t.order = order;
-
+
__trace_var(TRC_MEM_POD_POPULATE, 0, sizeof(t), &t);
}
@@ -1161,7 +1161,7 @@ remap_and_retry:
t.gfn = gfn;
t.d = d->domain_id;
-
+
__trace_var(TRC_MEM_POD_SUPERPAGE_SPLINTER, 0, sizeof(t), &t);
}
--
2.11.0
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xen.org
https://lists.xen.org/xen-devel
next prev parent reply other threads:[~2017-09-13 17:59 UTC|newest]
Thread overview: 24+ messages / expand[flat|nested] mbox.gz Atom feed top
2017-09-13 17:59 [PATCH 00/15] xen/x86: Clean-up the PoD code Julien Grall
2017-09-13 17:59 ` [PATCH 01/15] xen/x86: p2m-pod: Clean-up includes Julien Grall
2017-09-13 17:59 ` Julien Grall [this message]
2017-09-13 17:59 ` [PATCH 03/15] xen/x86: p2m-pod: Fix coding style for comments Julien Grall
2017-09-13 17:59 ` [PATCH 04/15] xen/x86: p2m-pod: Fix coding style Julien Grall
2017-09-13 17:59 ` [PATCH 05/15] xen/x86: p2m-pod: Avoid redundant assignments in p2m_pod_demand_populate Julien Grall
2017-09-13 17:59 ` [PATCH 06/15] xen/x86: p2m-pod: Clean-up use of typesafe MFN Julien Grall
2017-09-13 17:59 ` [PATCH 07/15] xen/x86: p2m-pod: Use typesafe gfn in p2m_pod_decrease_reservation Julien Grall
2017-09-13 17:59 ` [PATCH 08/15] xen/x86: p2m: Use typesafe gfn for the P2M callbacks get_entry and set_entry Julien Grall
2017-09-13 18:22 ` Andrew Cooper
2017-09-13 18:27 ` Julien Grall
2017-09-13 19:10 ` Razvan Cojocaru
2017-09-13 19:08 ` Razvan Cojocaru
2017-09-13 19:32 ` Julien Grall
2017-09-20 6:57 ` Tian, Kevin
2017-09-13 17:59 ` [PATCH 09/15] xen/x86: p2m: Use typesafe GFN in p2m_set_entry Julien Grall
2017-09-13 20:16 ` Tamas K Lengyel
2017-09-13 17:59 ` [PATCH 10/15] xen/x86: p2m-pod: Use typesafe GFN in pod_eager_record Julien Grall
2017-09-13 17:59 ` [PATCH 11/15] xen/x86: p2m-pod: Clean-up p2m_pod_zero_check Julien Grall
2017-09-13 17:59 ` [PATCH 12/15] xen/x86: p2m-pod: Use typesafe gfn in p2m_pod_zero_check Julien Grall
2017-09-13 17:59 ` [PATCH 13/15] xen/x86: p2m-pod: Use typesafe gfn in p2m_pod_demand_populate Julien Grall
2017-09-13 17:59 ` [PATCH 14/15] xen/x86: p2m-pod: Use typesafe gfn for the fields reclaim_single and max_guest Julien Grall
2017-09-13 17:59 ` [PATCH 15/15] xen/x86: p2m-pod: Rework prototype of p2m_pod_demand_populate Julien Grall
2017-09-13 18:30 ` [PATCH 00/15] xen/x86: Clean-up the PoD code Andrew Cooper
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20170913175953.16942-3-julien.grall@arm.com \
--to=julien.grall@arm.com \
--cc=andrew.cooper3@citrix.com \
--cc=george.dunlap@eu.citrix.com \
--cc=jbeulich@suse.com \
--cc=xen-devel@lists.xen.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).