aboutsummaryrefslogtreecommitdiff
path: root/migration.c
diff options
context:
space:
mode:
authorAlexey Kardashevskiy <aik@ozlabs.ru>2014-03-27 14:57:26 +1100
committerJuan Quintela <quintela@redhat.com>2014-06-16 04:55:27 +0200
commitf7cd55a023b71ff9583dad9135413b7cc5821dfe (patch)
tree7f75137f512e391b8d3e736d5c680806ca347d5a /migration.c
parentc6f6646c60a2450e437ee89582295d5fde8e8a72 (diff)
migration: Increase default max_downtime from 30ms to 300ms
The existing timeout is 30ms which on 100MB/s (1Gbit) gives us 3MB/s rate maximum. If we put some load on the guest, it is easy to get page dirtying rate too big so live migration will never complete. In the case of libvirt that means that the guest will be stopped anyway after a timeout specified in the "virsh migrate" command and this normally generates even bigger delay. This changes max_downtime to 300ms which seems to be more reasonable value. Signed-off-by: Alexey Kardashevskiy <aik@ozlabs.ru> Signed-off-by: Juan Quintela <quintela@redhat.com>
Diffstat (limited to 'migration.c')
-rw-r--r--migration.c2
1 files changed, 1 insertions, 1 deletions
diff --git a/migration.c b/migration.c
index 3fc03d6687..873fa96e54 100644
--- a/migration.c
+++ b/migration.c
@@ -133,7 +133,7 @@ void process_incoming_migration(QEMUFile *f)
* the choice of nanoseconds is because it is the maximum resolution that
* get_clock() can achieve. It is an internal measure. All user-visible
* units must be in seconds */
-static uint64_t max_downtime = 30000000;
+static uint64_t max_downtime = 300000000;
uint64_t migrate_max_downtime(void)
{