All of lore.kernel.org
 help / color / mirror / Atom feed
From: Marcelo Tosatti <mtosatti@redhat.com>
To: kvm@vger.kernel.org
Cc: johnstul@us.ibm.com, jeremy@goop.org, glommer@parallels.com,
	zamsden@gmail.com, gleb@redhat.com, avi@redhat.com,
	pbonzini@redhat.com, Marcelo Tosatti <mtosatti@redhat.com>
Subject: [patch 17/18] KVM: x86: require matched TSC offsets for master clock
Date: Wed, 14 Nov 2012 22:08:40 -0200	[thread overview]
Message-ID: <20121115000944.724933468@redhat.com> (raw)
In-Reply-To: 20121115000823.285102321@redhat.com

[-- Attachment #1: 17-masterclock-require-matched-tsc --]
[-- Type: text/plain, Size: 6962 bytes --]

With master clock, a pvclock clock read calculates:

ret = system_timestamp + [ (rdtsc + tsc_offset) - tsc_timestamp ]

Where 'rdtsc' is the host TSC.

system_timestamp and tsc_timestamp are unique, one tuple 
per VM: the "master clock".

Given a host with synchronized TSCs, its obvious that
guest TSC must be matched for the above to guarantee monotonicity.

Allow master clock usage only if guest TSCs are synchronized.

Signed-off-by: Marcelo Tosatti <mtosatti@redhat.com>

Index: vsyscall/arch/x86/include/asm/kvm_host.h
===================================================================
--- vsyscall.orig/arch/x86/include/asm/kvm_host.h
+++ vsyscall/arch/x86/include/asm/kvm_host.h
@@ -560,6 +560,7 @@ struct kvm_arch {
 	u64 cur_tsc_write;
 	u64 cur_tsc_offset;
 	u8  cur_tsc_generation;
+	int nr_vcpus_matched_tsc;
 
 	spinlock_t pvclock_gtod_sync_lock;
 	bool use_master_clock;
Index: vsyscall/arch/x86/kvm/x86.c
===================================================================
--- vsyscall.orig/arch/x86/kvm/x86.c
+++ vsyscall/arch/x86/kvm/x86.c
@@ -1097,12 +1097,38 @@ static u64 compute_guest_tsc(struct kvm_
 	return tsc;
 }
 
+void kvm_track_tsc_matching(struct kvm_vcpu *vcpu)
+{
+	bool vcpus_matched;
+	bool do_request = false;
+	struct kvm_arch *ka = &vcpu->kvm->arch;
+	struct pvclock_gtod_data *gtod = &pvclock_gtod_data;
+
+	vcpus_matched = (ka->nr_vcpus_matched_tsc + 1 ==
+			 atomic_read(&vcpu->kvm->online_vcpus));
+
+	if (vcpus_matched && gtod->clock.vclock_mode == VCLOCK_TSC)
+		if (!ka->use_master_clock)
+			do_request = 1;
+
+	if (!vcpus_matched && ka->use_master_clock)
+			do_request = 1;
+
+	if (do_request)
+		kvm_make_request(KVM_REQ_MASTERCLOCK_UPDATE, vcpu);
+
+	trace_kvm_track_tsc(vcpu->vcpu_id, ka->nr_vcpus_matched_tsc,
+			    atomic_read(&vcpu->kvm->online_vcpus),
+		            ka->use_master_clock, gtod->clock.vclock_mode);
+}
+
 void kvm_write_tsc(struct kvm_vcpu *vcpu, u64 data)
 {
 	struct kvm *kvm = vcpu->kvm;
 	u64 offset, ns, elapsed;
 	unsigned long flags;
 	s64 usdiff;
+	bool matched;
 
 	raw_spin_lock_irqsave(&kvm->arch.tsc_write_lock, flags);
 	offset = kvm_x86_ops->compute_tsc_offset(vcpu, data);
@@ -1145,6 +1171,7 @@ void kvm_write_tsc(struct kvm_vcpu *vcpu
 			offset = kvm_x86_ops->compute_tsc_offset(vcpu, data);
 			pr_debug("kvm: adjusted tsc offset by %llu\n", delta);
 		}
+		matched = true;
 	} else {
 		/*
 		 * We split periods of matched TSC writes into generations.
@@ -1159,6 +1186,7 @@ void kvm_write_tsc(struct kvm_vcpu *vcpu
 		kvm->arch.cur_tsc_nsec = ns;
 		kvm->arch.cur_tsc_write = data;
 		kvm->arch.cur_tsc_offset = offset;
+		matched = false;
 		pr_debug("kvm: new tsc generation %u, clock %llu\n",
 			 kvm->arch.cur_tsc_generation, data);
 	}
@@ -1182,6 +1210,15 @@ void kvm_write_tsc(struct kvm_vcpu *vcpu
 
 	kvm_x86_ops->write_tsc_offset(vcpu, offset);
 	raw_spin_unlock_irqrestore(&kvm->arch.tsc_write_lock, flags);
+
+	spin_lock(&kvm->arch.pvclock_gtod_sync_lock);
+	if (matched)
+		kvm->arch.nr_vcpus_matched_tsc++;
+	else
+		kvm->arch.nr_vcpus_matched_tsc = 0;
+
+	kvm_track_tsc_matching(vcpu);
+	spin_unlock(&kvm->arch.pvclock_gtod_sync_lock);
 }
 
 EXPORT_SYMBOL_GPL(kvm_write_tsc);
@@ -1271,8 +1308,9 @@ static bool kvm_get_time_and_clockread(s
 
 /*
  *
- * Assuming a stable TSC across physical CPUS, the following condition
- * is possible. Each numbered line represents an event visible to both
+ * Assuming a stable TSC across physical CPUS, and a stable TSC
+ * across virtual CPUs, the following condition is possible.
+ * Each numbered line represents an event visible to both
  * CPUs at the next numbered event.
  *
  * "timespecX" represents host monotonic time. "tscX" represents
@@ -1305,7 +1343,7 @@ static bool kvm_get_time_and_clockread(s
  * copy of host monotonic time values. Update that master copy
  * in lockstep.
  *
- * Rely on synchronization of host TSCs for monotonicity.
+ * Rely on synchronization of host TSCs and guest TSCs for monotonicity.
  *
  */
 
@@ -1313,20 +1351,27 @@ static void pvclock_update_vm_gtod_copy(
 {
 	struct kvm_arch *ka = &kvm->arch;
 	int vclock_mode;
+	bool host_tsc_clocksource, vcpus_matched;
+
+	vcpus_matched = (ka->nr_vcpus_matched_tsc + 1 ==
+				atomic_read(&kvm->online_vcpus));
 
 	/*
  	 * If the host uses TSC clock, then passthrough TSC as stable
 	 * to the guest.
 	 */
-	ka->use_master_clock = kvm_get_time_and_clockread(
+	host_tsc_clocksource = kvm_get_time_and_clockread(
 					&ka->master_kernel_ns,
 					&ka->master_cycle_now);
 
+	ka->use_master_clock = host_tsc_clocksource & vcpus_matched;
+
 	if (ka->use_master_clock)
 		atomic_set(&kvm_guest_has_master_clock, 1);
 
 	vclock_mode = pvclock_gtod_data.clock.vclock_mode;
-	trace_kvm_update_master_clock(ka->use_master_clock, vclock_mode);
+	trace_kvm_update_master_clock(ka->use_master_clock, vclock_mode,
+				      vcpus_matched);
 }
 
 static int kvm_guest_time_update(struct kvm_vcpu *v)
Index: vsyscall/arch/x86/kvm/trace.h
===================================================================
--- vsyscall.orig/arch/x86/kvm/trace.h
+++ vsyscall/arch/x86/kvm/trace.h
@@ -762,21 +762,54 @@ TRACE_EVENT(
 	{VCLOCK_HPET, "hpet"}			\
 
 TRACE_EVENT(kvm_update_master_clock,
-	TP_PROTO(bool use_master_clock, unsigned int host_clock),
-	TP_ARGS(use_master_clock, host_clock),
+	TP_PROTO(bool use_master_clock, unsigned int host_clock, bool offset_matched),
+	TP_ARGS(use_master_clock, host_clock, offset_matched),
 
 	TP_STRUCT__entry(
 		__field(		bool,	use_master_clock	)
 		__field(	unsigned int,	host_clock		)
+		__field(		bool,	offset_matched		)
 	),
 
 	TP_fast_assign(
 		__entry->use_master_clock	= use_master_clock;
 		__entry->host_clock		= host_clock;
+		__entry->offset_matched		= offset_matched;
 	),
 
-	TP_printk("masterclock %d hostclock %s",
+	TP_printk("masterclock %d hostclock %s offsetmatched %u",
 		  __entry->use_master_clock,
+		  __print_symbolic(__entry->host_clock, host_clocks),
+		  __entry->offset_matched)
+);
+
+TRACE_EVENT(kvm_track_tsc,
+	TP_PROTO(unsigned int vcpu_id, unsigned int nr_matched,
+		 unsigned int online_vcpus, bool use_master_clock,
+		 unsigned int host_clock),
+	TP_ARGS(vcpu_id, nr_matched, online_vcpus, use_master_clock,
+		host_clock),
+
+	TP_STRUCT__entry(
+		__field(	unsigned int,	vcpu_id			)
+		__field(	unsigned int,	nr_vcpus_matched_tsc	)
+		__field(	unsigned int,	online_vcpus		)
+		__field(	bool,		use_master_clock	)
+		__field(	unsigned int,	host_clock		)
+	),
+
+	TP_fast_assign(
+		__entry->vcpu_id		= vcpu_id;
+		__entry->nr_vcpus_matched_tsc	= nr_matched;
+		__entry->online_vcpus		= online_vcpus;
+		__entry->use_master_clock	= use_master_clock;
+		__entry->host_clock		= host_clock;
+	),
+
+	TP_printk("vcpu_id %u masterclock %u offsetmatched %u nr_online %u"
+		  " hostclock %s",
+		  __entry->vcpu_id, __entry->use_master_clock,
+		  __entry->nr_vcpus_matched_tsc, __entry->online_vcpus,
 		  __print_symbolic(__entry->host_clock, host_clocks))
 );
 



  parent reply	other threads:[~2012-11-15  1:27 UTC|newest]

Thread overview: 49+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2012-11-15  0:08 [patch 00/18] pvclock vsyscall support + KVM hypervisor support (v4) Marcelo Tosatti
2012-11-15  0:08 ` [patch 01/18] KVM: x86: retain pvclock guest stopped bit in guest memory Marcelo Tosatti
2012-11-15  0:08 ` [patch 02/18] x86: kvmclock: allocate pvclock shared memory area Marcelo Tosatti
2012-11-15 17:05   ` Glauber Costa
2012-11-16  2:07     ` [patch 02/18] x86: kvmclock: allocate pvclock shared memory area (v2) Marcelo Tosatti
2012-11-16  7:06       ` Glauber Costa
2012-11-15  0:08 ` [patch 03/18] x86: pvclock: make sure rdtsc doesnt speculate out of region Marcelo Tosatti
2012-11-15  0:08 ` [patch 04/18] x86: pvclock: remove pvclock_shadow_time Marcelo Tosatti
2012-11-15  0:08 ` [patch 05/18] x86: pvclock: create helper for pvclock data retrieval Marcelo Tosatti
2012-11-15 12:27   ` Glauber Costa
2012-11-15  0:08 ` [patch 06/18] x86: pvclock: introduce helper to read flags Marcelo Tosatti
2012-11-15 12:28   ` Glauber Costa
2012-11-15  0:08 ` [patch 07/18] x86: pvclock: add note about rdtsc barriers Marcelo Tosatti
2012-11-15 12:30   ` Glauber Costa
2012-11-16  2:05     ` [patch 07/18] x86: pvclock: add note about rdtsc barriers (v2) Marcelo Tosatti
2012-11-15  0:08 ` [patch 08/18] sched: add notifier for cross-cpu migrations Marcelo Tosatti
2012-11-15  7:01   ` Gleb Natapov
2012-11-15  0:08 ` [patch 09/18] x86: pvclock: generic pvclock vsyscall initialization Marcelo Tosatti
2012-11-15  0:08 ` [patch 10/18] x86: kvm guest: pvclock vsyscall support Marcelo Tosatti
2012-11-15  0:08 ` [patch 11/18] x86: vdso: pvclock gettime support Marcelo Tosatti
2012-11-15  0:08 ` [patch 12/18] KVM: x86: pass host_tsc to read_l1_tsc Marcelo Tosatti
2012-11-15  0:08 ` [patch 13/18] time: export time information for KVM pvclock Marcelo Tosatti
2012-11-15  1:38   ` John Stultz
2012-11-15  0:08 ` [patch 14/18] KVM: x86: notifier for clocksource changes Marcelo Tosatti
2012-11-15  0:08 ` [patch 15/18] KVM: x86: implement PVCLOCK_TSC_STABLE_BIT pvclock flag Marcelo Tosatti
2012-11-15  0:08 ` [patch 16/18] KVM: x86: add kvm_arch_vcpu_postcreate callback, move TSC initialization Marcelo Tosatti
2012-11-15  0:08 ` Marcelo Tosatti [this message]
2012-11-15  0:08 ` [patch 18/18] KVM: x86: update pvclock area conditionally, on cpu migration Marcelo Tosatti
2012-11-15 12:34   ` Glauber Costa
2012-11-19 21:57 ` [patch 00/18] pvclock vsyscall support + KVM hypervisor support (v5) Marcelo Tosatti
2012-11-19 21:57   ` [patch 01/18] KVM: x86: retain pvclock guest stopped bit in guest memory Marcelo Tosatti
2012-11-19 21:58   ` [patch 02/18] x86: kvmclock: allocate pvclock shared memory area Marcelo Tosatti
2012-11-19 21:58   ` [patch 03/18] x86: pvclock: make sure rdtsc doesnt speculate out of region Marcelo Tosatti
2012-11-19 21:58   ` [patch 04/18] x86: pvclock: remove pvclock_shadow_time Marcelo Tosatti
2012-11-19 21:58   ` [patch 05/18] x86: pvclock: create helper for pvclock data retrieval Marcelo Tosatti
2012-11-19 21:58   ` [patch 06/18] x86: pvclock: introduce helper to read flags Marcelo Tosatti
2012-11-19 21:58   ` [patch 07/18] x86: pvclock: add note about rdtsc barriers Marcelo Tosatti
2012-11-19 21:58   ` [patch 08/18] sched: add notifier for cross-cpu migrations Marcelo Tosatti
2012-11-19 21:58   ` [patch 09/18] x86: pvclock: generic pvclock vsyscall initialization Marcelo Tosatti
2012-11-19 21:58   ` [patch 11/18] x86: vdso: pvclock gettime support Marcelo Tosatti
2012-11-19 21:58   ` [patch 12/18] KVM: x86: pass host_tsc to read_l1_tsc Marcelo Tosatti
2012-11-19 21:58   ` [patch 13/18] time: export time information for KVM pvclock Marcelo Tosatti
2012-11-19 21:58   ` [patch 14/18] KVM: x86: notifier for clocksource changes Marcelo Tosatti
2012-11-19 21:58   ` [patch 15/18] KVM: x86: implement PVCLOCK_TSC_STABLE_BIT pvclock flag Marcelo Tosatti
2012-11-19 21:58   ` [patch 16/18] KVM: x86: add kvm_arch_vcpu_postcreate callback, move TSC initialization Marcelo Tosatti
2012-11-19 21:58   ` [patch 17/18] KVM: x86: require matched TSC offsets for master clock Marcelo Tosatti
2012-11-19 21:58   ` [patch 18/18] KVM: x86: update pvclock area conditionally, on cpu migration Marcelo Tosatti
2012-11-20  9:01     ` Glauber Costa
2012-11-20  9:03   ` [patch 00/18] pvclock vsyscall support + KVM hypervisor support (v5) Glauber Costa

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20121115000944.724933468@redhat.com \
    --to=mtosatti@redhat.com \
    --cc=avi@redhat.com \
    --cc=gleb@redhat.com \
    --cc=glommer@parallels.com \
    --cc=jeremy@goop.org \
    --cc=johnstul@us.ibm.com \
    --cc=kvm@vger.kernel.org \
    --cc=pbonzini@redhat.com \
    --cc=zamsden@gmail.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.