qemu-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [RFC] migration: Remove old compression code


From: Dr. David Alan Gilbert
Subject: Re: [RFC] migration: Remove old compression code
Date: Thu, 23 Jan 2020 12:22:21 +0000
User-agent: Mutt/1.13.0 (2019-11-30)

* Juan Quintela (address@hidden) wrote:
> Hi

Lets cc in some people who have previously contributed to the existing
compression code.

> There are several problems with the old compression code:
> * it compress one page at a time (i.e. 4KB)
> 
> * it has to copy all the page to the compression thread, the
>   compression thread compress it, it copies back the compressed data
>   to the main thread and it sends it there (equivalent for decompress).
> 
> * It is really slow (both the setup of it, and the compression of each page)
> 
> You can see that in the previous series, I added compression to multifd, its 
> advantages are:
> * We compress "chunks" of 128 pages at a time
> * We don't copy anything, we just compress everything that is sent through 
> the channel
> * We don't send data over the main channel, so we have less overhead
> * I added zstd support, that
>   o  Compress much more
>   o  It is way faster
> 
> * Using the migration-test values (best option for compressors, as we
>   only change one byte on each page):
>   * zstd is around the same latency than no compression (i.e. it takes the 
> same time)
>   * zlib needs 2500 bytes to send 128 pages compressed (remember 1
>     byte changed in each page, rest are zeros)
>   * zstd needs 52 bytes for the same packets (yes, I checked several
>     times, it appears that its use of dictionaries is good here)
> 
> This is an RFC, you will ask why?
> There are several questions:
> * How to proceed to remove this?
>    - Deprecate it into the future?
>      Has the advantage that customers can continue using it.
>      But it is not so stable to use it.
>    - Remove it now?
>      Has the advantage of much less code, but ....
> * Statistics:
>   Several of the statistics are already only printed when compression is
>   being used. My understanding is that we ase safe removing this ones.
> * But whan about the migration_parameters?  Should we leave them
>   alone, and just don't care about them? Should we remove them?
> * What to do with migration.json changes?  I will apprecciate comments here.
> 
> Thanks, Juan.
> 
> PD.  You can see from the previous series that putting compression on
>      top of multifd + splitting the code in different files + making it 
> optional:
>      around 1040 new lines of code.
>      Removing this ~ 960 lines of code.
> 
> Signed-off-by: Juan Quintela <address@hidden>
> ---
>  hw/core/machine.c     |   1 -
>  migration/migration.c | 112 --------
>  migration/migration.h |  11 -
>  migration/qemu-file.c |  77 -----
>  migration/qemu-file.h |   3 -
>  migration/ram.c       | 650 +-----------------------------------------
>  migration/ram.h       |   1 -
>  monitor/hmp-cmds.c    |  37 ---
>  qapi/migration.json   |  79 +----
>  9 files changed, 9 insertions(+), 962 deletions(-)
> 
> diff --git a/hw/core/machine.c b/hw/core/machine.c
> index 3e288bfceb..2424662f5e 100644
> --- a/hw/core/machine.c
> +++ b/hw/core/machine.c
> @@ -76,7 +76,6 @@ GlobalProperty hw_compat_3_0[] = {};
>  const size_t hw_compat_3_0_len = G_N_ELEMENTS(hw_compat_3_0);
>  
>  GlobalProperty hw_compat_2_12[] = {
> -    { "migration", "decompress-error-check", "off" },
>      { "hda-audio", "use-timer", "false" },
>      { "cirrus-vga", "global-vmstate", "true" },
>      { "VGA", "global-vmstate", "true" },
> diff --git a/migration/migration.c b/migration/migration.c
> index f0d5ade1df..fad5c4faf4 100644
> --- a/migration/migration.c
> +++ b/migration/migration.c
> @@ -70,11 +70,6 @@
>  #define MAX_MIGRATE_DOWNTIME_SECONDS 2000
>  #define MAX_MIGRATE_DOWNTIME (MAX_MIGRATE_DOWNTIME_SECONDS * 1000)
>  
> -/* Default compression thread count */
> -#define DEFAULT_MIGRATE_COMPRESS_THREAD_COUNT 8
> -/* Default decompression thread count, usually decompression is at
> - * least 4 times as fast as compression.*/
> -#define DEFAULT_MIGRATE_DECOMPRESS_THREAD_COUNT 2
>  /*0: means nocompress, 1: best speed, ... 9: best compress ratio */
>  #define DEFAULT_MIGRATE_COMPRESS_LEVEL 1
>  /* Define default autoconverge cpu throttle migration parameters */
> @@ -773,12 +768,6 @@ MigrationParameters *qmp_query_migrate_parameters(Error 
> **errp)
>      params = g_malloc0(sizeof(*params));
>      params->has_compress_level = true;
>      params->compress_level = s->parameters.compress_level;
> -    params->has_compress_threads = true;
> -    params->compress_threads = s->parameters.compress_threads;
> -    params->has_compress_wait_thread = true;
> -    params->compress_wait_thread = s->parameters.compress_wait_thread;
> -    params->has_decompress_threads = true;
> -    params->decompress_threads = s->parameters.decompress_threads;
>      params->has_cpu_throttle_initial = true;
>      params->cpu_throttle_initial = s->parameters.cpu_throttle_initial;
>      params->has_cpu_throttle_increment = true;
> @@ -926,18 +915,6 @@ static void populate_ram_info(MigrationInfo *info, 
> MigrationState *s)
>          info->xbzrle_cache->overflow = xbzrle_counters.overflow;
>      }
>  
> -    if (migrate_use_compression()) {
> -        info->has_compression = true;
> -        info->compression = g_malloc0(sizeof(*info->compression));
> -        info->compression->pages = compression_counters.pages;
> -        info->compression->busy = compression_counters.busy;
> -        info->compression->busy_rate = compression_counters.busy_rate;
> -        info->compression->compressed_size =
> -                                    compression_counters.compressed_size;
> -        info->compression->compression_rate =
> -                                    compression_counters.compression_rate;
> -    }
> -
>      if (cpu_throttle_active()) {
>          info->has_cpu_throttle_percentage = true;
>          info->cpu_throttle_percentage = cpu_throttle_get_percentage();
> @@ -1153,20 +1130,6 @@ static bool migrate_params_check(MigrationParameters 
> *params, Error **errp)
>          return false;
>      }
>  
> -    if (params->has_compress_threads && (params->compress_threads < 1)) {
> -        error_setg(errp, QERR_INVALID_PARAMETER_VALUE,
> -                   "compress_threads",
> -                   "is invalid, it should be in the range of 1 to 255");
> -        return false;
> -    }
> -
> -    if (params->has_decompress_threads && (params->decompress_threads < 1)) {
> -        error_setg(errp, QERR_INVALID_PARAMETER_VALUE,
> -                   "decompress_threads",
> -                   "is invalid, it should be in the range of 1 to 255");
> -        return false;
> -    }
> -
>      if (params->has_cpu_throttle_initial &&
>          (params->cpu_throttle_initial < 1 ||
>           params->cpu_throttle_initial > 99)) {
> @@ -1276,18 +1239,6 @@ static void 
> migrate_params_test_apply(MigrateSetParameters *params,
>          dest->compress_level = params->compress_level;
>      }
>  
> -    if (params->has_compress_threads) {
> -        dest->compress_threads = params->compress_threads;
> -    }
> -
> -    if (params->has_compress_wait_thread) {
> -        dest->compress_wait_thread = params->compress_wait_thread;
> -    }
> -
> -    if (params->has_decompress_threads) {
> -        dest->decompress_threads = params->decompress_threads;
> -    }
> -
>      if (params->has_cpu_throttle_initial) {
>          dest->cpu_throttle_initial = params->cpu_throttle_initial;
>      }
> @@ -1360,18 +1311,6 @@ static void migrate_params_apply(MigrateSetParameters 
> *params, Error **errp)
>          s->parameters.compress_level = params->compress_level;
>      }
>  
> -    if (params->has_compress_threads) {
> -        s->parameters.compress_threads = params->compress_threads;
> -    }
> -
> -    if (params->has_compress_wait_thread) {
> -        s->parameters.compress_wait_thread = params->compress_wait_thread;
> -    }
> -
> -    if (params->has_decompress_threads) {
> -        s->parameters.decompress_threads = params->decompress_threads;
> -    }
> -
>      if (params->has_cpu_throttle_initial) {
>          s->parameters.cpu_throttle_initial = params->cpu_throttle_initial;
>      }
> @@ -2142,15 +2081,6 @@ bool migrate_postcopy_blocktime(void)
>      return s->enabled_capabilities[MIGRATION_CAPABILITY_POSTCOPY_BLOCKTIME];
>  }
>  
> -bool migrate_use_compression(void)
> -{
> -    MigrationState *s;
> -
> -    s = migrate_get_current();
> -
> -    return s->enabled_capabilities[MIGRATION_CAPABILITY_COMPRESS];
> -}
> -
>  int migrate_compress_level(void)
>  {
>      MigrationState *s;
> @@ -2160,33 +2090,6 @@ int migrate_compress_level(void)
>      return s->parameters.compress_level;
>  }
>  
> -int migrate_compress_threads(void)
> -{
> -    MigrationState *s;
> -
> -    s = migrate_get_current();
> -
> -    return s->parameters.compress_threads;
> -}
> -
> -int migrate_compress_wait_thread(void)
> -{
> -    MigrationState *s;
> -
> -    s = migrate_get_current();
> -
> -    return s->parameters.compress_wait_thread;
> -}
> -
> -int migrate_decompress_threads(void)
> -{
> -    MigrationState *s;
> -
> -    s = migrate_get_current();
> -
> -    return s->parameters.decompress_threads;
> -}
> -
>  bool migrate_dirty_bitmaps(void)
>  {
>      MigrationState *s;
> @@ -3489,8 +3392,6 @@ void migration_global_dump(Monitor *mon)
>                     ms->send_configuration ? "on" : "off");
>      monitor_printf(mon, "send-section-footer: %s\n",
>                     ms->send_section_footer ? "on" : "off");
> -    monitor_printf(mon, "decompress-error-check: %s\n",
> -                   ms->decompress_error_check ? "on" : "off");
>      monitor_printf(mon, "clear-bitmap-shift: %u\n",
>                     ms->clear_bitmap_shift);
>  }
> @@ -3505,8 +3406,6 @@ static Property migration_properties[] = {
>                       send_configuration, true),
>      DEFINE_PROP_BOOL("send-section-footer", MigrationState,
>                       send_section_footer, true),
> -    DEFINE_PROP_BOOL("decompress-error-check", MigrationState,
> -                      decompress_error_check, true),
>      DEFINE_PROP_UINT8("x-clear-bitmap-shift", MigrationState,
>                        clear_bitmap_shift, CLEAR_BITMAP_SHIFT_DEFAULT),
>  
> @@ -3514,14 +3413,6 @@ static Property migration_properties[] = {
>      DEFINE_PROP_UINT8("x-compress-level", MigrationState,
>                        parameters.compress_level,
>                        DEFAULT_MIGRATE_COMPRESS_LEVEL),
> -    DEFINE_PROP_UINT8("x-compress-threads", MigrationState,
> -                      parameters.compress_threads,
> -                      DEFAULT_MIGRATE_COMPRESS_THREAD_COUNT),
> -    DEFINE_PROP_BOOL("x-compress-wait-thread", MigrationState,
> -                      parameters.compress_wait_thread, true),
> -    DEFINE_PROP_UINT8("x-decompress-threads", MigrationState,
> -                      parameters.decompress_threads,
> -                      DEFAULT_MIGRATE_DECOMPRESS_THREAD_COUNT),
>      DEFINE_PROP_UINT8("x-cpu-throttle-initial", MigrationState,
>                        parameters.cpu_throttle_initial,
>                        DEFAULT_MIGRATE_CPU_THROTTLE_INITIAL),
> @@ -3569,7 +3460,6 @@ static Property migration_properties[] = {
>      DEFINE_PROP_MIG_CAP("x-rdma-pin-all", MIGRATION_CAPABILITY_RDMA_PIN_ALL),
>      DEFINE_PROP_MIG_CAP("x-auto-converge", 
> MIGRATION_CAPABILITY_AUTO_CONVERGE),
>      DEFINE_PROP_MIG_CAP("x-zero-blocks", MIGRATION_CAPABILITY_ZERO_BLOCKS),
> -    DEFINE_PROP_MIG_CAP("x-compress", MIGRATION_CAPABILITY_COMPRESS),
>      DEFINE_PROP_MIG_CAP("x-events", MIGRATION_CAPABILITY_EVENTS),
>      DEFINE_PROP_MIG_CAP("x-postcopy-ram", MIGRATION_CAPABILITY_POSTCOPY_RAM),
>      DEFINE_PROP_MIG_CAP("x-colo", MIGRATION_CAPABILITY_X_COLO),
> @@ -3623,8 +3513,6 @@ static void migration_instance_init(Object *obj)
>  
>      /* Set has_* up only for parameter checks */
>      params->has_compress_level = true;
> -    params->has_compress_threads = true;
> -    params->has_decompress_threads = true;
>      params->has_cpu_throttle_initial = true;
>      params->has_cpu_throttle_increment = true;
>      params->has_max_bandwidth = true;
> diff --git a/migration/migration.h b/migration/migration.h
> index 3b0b413a93..21ef1838d2 100644
> --- a/migration/migration.h
> +++ b/migration/migration.h
> @@ -244,13 +244,6 @@ struct MigrationState
>      /* Needed by postcopy-pause state */
>      QemuSemaphore postcopy_pause_sem;
>      QemuSemaphore postcopy_pause_rp_sem;
> -    /*
> -     * Whether we abort the migration if decompression errors are
> -     * detected at the destination. It is left at false for qemu
> -     * older than 3.0, since only newer qemu sends streams that
> -     * do not trigger spurious decompression errors.
> -     */
> -    bool decompress_error_check;
>  
>      /*
>       * This decides the size of guest memory chunk that will be used
> @@ -313,11 +306,7 @@ bool migrate_use_return_path(void);
>  
>  uint64_t ram_get_total_transferred_pages(void);
>  
> -bool migrate_use_compression(void);
>  int migrate_compress_level(void);
> -int migrate_compress_threads(void);
> -int migrate_compress_wait_thread(void);
> -int migrate_decompress_threads(void);
>  bool migrate_use_events(void);
>  bool migrate_postcopy_blocktime(void);
>  
> diff --git a/migration/qemu-file.c b/migration/qemu-file.c
> index bbb2b63927..1f14744f22 100644
> --- a/migration/qemu-file.c
> +++ b/migration/qemu-file.c
> @@ -737,83 +737,6 @@ uint64_t qemu_get_be64(QEMUFile *f)
>      return v;
>  }
>  
> -/* return the size after compression, or negative value on error */
> -static int qemu_compress_data(z_stream *stream, uint8_t *dest, size_t 
> dest_len,
> -                              const uint8_t *source, size_t source_len)
> -{
> -    int err;
> -
> -    err = deflateReset(stream);
> -    if (err != Z_OK) {
> -        return -1;
> -    }
> -
> -    stream->avail_in = source_len;
> -    stream->next_in = (uint8_t *)source;
> -    stream->avail_out = dest_len;
> -    stream->next_out = dest;
> -
> -    err = deflate(stream, Z_FINISH);
> -    if (err != Z_STREAM_END) {
> -        return -1;
> -    }
> -
> -    return stream->next_out - dest;
> -}
> -
> -/* Compress size bytes of data start at p and store the compressed
> - * data to the buffer of f.
> - *
> - * When f is not writable, return -1 if f has no space to save the
> - * compressed data.
> - * When f is wirtable and it has no space to save the compressed data,
> - * do fflush first, if f still has no space to save the compressed
> - * data, return -1.
> - */
> -ssize_t qemu_put_compression_data(QEMUFile *f, z_stream *stream,
> -                                  const uint8_t *p, size_t size)
> -{
> -    ssize_t blen = IO_BUF_SIZE - f->buf_index - sizeof(int32_t);
> -
> -    if (blen < compressBound(size)) {
> -        if (!qemu_file_is_writable(f)) {
> -            return -1;
> -        }
> -        qemu_fflush(f);
> -        blen = IO_BUF_SIZE - sizeof(int32_t);
> -        if (blen < compressBound(size)) {
> -            return -1;
> -        }
> -    }
> -
> -    blen = qemu_compress_data(stream, f->buf + f->buf_index + 
> sizeof(int32_t),
> -                              blen, p, size);
> -    if (blen < 0) {
> -        return -1;
> -    }
> -
> -    qemu_put_be32(f, blen);
> -    add_buf_to_iovec(f, blen);
> -    return blen + sizeof(int32_t);
> -}
> -
> -/* Put the data in the buffer of f_src to the buffer of f_des, and
> - * then reset the buf_index of f_src to 0.
> - */
> -
> -int qemu_put_qemu_file(QEMUFile *f_des, QEMUFile *f_src)
> -{
> -    int len = 0;
> -
> -    if (f_src->buf_index > 0) {
> -        len = f_src->buf_index;
> -        qemu_put_buffer(f_des, f_src->buf, f_src->buf_index);
> -        f_src->buf_index = 0;
> -        f_src->iovcnt = 0;
> -    }
> -    return len;
> -}
> -
>  /*
>   * Get a string whose length is determined by a single preceding byte
>   * A preallocated 256 byte buffer must be passed in.
> diff --git a/migration/qemu-file.h b/migration/qemu-file.h
> index a9b6d6ccb7..dd0161fbdc 100644
> --- a/migration/qemu-file.h
> +++ b/migration/qemu-file.h
> @@ -138,9 +138,6 @@ bool qemu_file_is_writable(QEMUFile *f);
>  
>  size_t qemu_peek_buffer(QEMUFile *f, uint8_t **buf, size_t size, size_t 
> offset);
>  size_t qemu_get_buffer_in_place(QEMUFile *f, uint8_t **buf, size_t size);
> -ssize_t qemu_put_compression_data(QEMUFile *f, z_stream *stream,
> -                                  const uint8_t *p, size_t size);
> -int qemu_put_qemu_file(QEMUFile *f_des, QEMUFile *f_src);
>  
>  /*
>   * Note that you can only peek continuous bytes from where the current 
> pointer
> diff --git a/migration/ram.c b/migration/ram.c
> index 0ef68798d2..557dbb104b 100644
> --- a/migration/ram.c
> +++ b/migration/ram.c
> @@ -73,7 +73,7 @@
>  #define RAM_SAVE_FLAG_CONTINUE 0x20
>  #define RAM_SAVE_FLAG_XBZRLE   0x40
>  /* 0x80 is reserved in migration.h start with 0x100 next */
> -#define RAM_SAVE_FLAG_COMPRESS_PAGE    0x100
> +/* 0x100 was RAM_SAVE_FLAG_COMPRESS_PAGE */
>  
>  static inline bool is_zero_range(uint8_t *p, uint64_t size)
>  {
> @@ -327,15 +327,6 @@ struct RAMState {
>      uint64_t num_dirty_pages_period;
>      /* xbzrle misses since the beginning of the period */
>      uint64_t xbzrle_cache_miss_prev;
> -
> -    /* compression statistics since the beginning of the period */
> -    /* amount of count that no free thread to compress data */
> -    uint64_t compress_thread_busy_prev;
> -    /* amount bytes after compression */
> -    uint64_t compressed_size_prev;
> -    /* amount of compressed pages */
> -    uint64_t compress_pages_prev;
> -
>      /* total handled target pages at the beginning of period */
>      uint64_t target_page_count_prev;
>      /* total handled target pages since start */
> @@ -408,172 +399,6 @@ struct PageSearchStatus {
>  };
>  typedef struct PageSearchStatus PageSearchStatus;
>  
> -CompressionStats compression_counters;
> -
> -struct CompressParam {
> -    bool done;
> -    bool quit;
> -    bool zero_page;
> -    QEMUFile *file;
> -    QemuMutex mutex;
> -    QemuCond cond;
> -    RAMBlock *block;
> -    ram_addr_t offset;
> -
> -    /* internally used fields */
> -    z_stream stream;
> -    uint8_t *originbuf;
> -};
> -typedef struct CompressParam CompressParam;
> -
> -struct DecompressParam {
> -    bool done;
> -    bool quit;
> -    QemuMutex mutex;
> -    QemuCond cond;
> -    void *des;
> -    uint8_t *compbuf;
> -    int len;
> -    z_stream stream;
> -};
> -typedef struct DecompressParam DecompressParam;
> -
> -static CompressParam *comp_param;
> -static QemuThread *compress_threads;
> -/* comp_done_cond is used to wake up the migration thread when
> - * one of the compression threads has finished the compression.
> - * comp_done_lock is used to co-work with comp_done_cond.
> - */
> -static QemuMutex comp_done_lock;
> -static QemuCond comp_done_cond;
> -/* The empty QEMUFileOps will be used by file in CompressParam */
> -static const QEMUFileOps empty_ops = { };
> -
> -static QEMUFile *decomp_file;
> -static DecompressParam *decomp_param;
> -static QemuThread *decompress_threads;
> -static QemuMutex decomp_done_lock;
> -static QemuCond decomp_done_cond;
> -
> -static bool do_compress_ram_page(QEMUFile *f, z_stream *stream, RAMBlock 
> *block,
> -                                 ram_addr_t offset, uint8_t *source_buf);
> -
> -static void *do_data_compress(void *opaque)
> -{
> -    CompressParam *param = opaque;
> -    RAMBlock *block;
> -    ram_addr_t offset;
> -    bool zero_page;
> -
> -    qemu_mutex_lock(&param->mutex);
> -    while (!param->quit) {
> -        if (param->block) {
> -            block = param->block;
> -            offset = param->offset;
> -            param->block = NULL;
> -            qemu_mutex_unlock(&param->mutex);
> -
> -            zero_page = do_compress_ram_page(param->file, &param->stream,
> -                                             block, offset, 
> param->originbuf);
> -
> -            qemu_mutex_lock(&comp_done_lock);
> -            param->done = true;
> -            param->zero_page = zero_page;
> -            qemu_cond_signal(&comp_done_cond);
> -            qemu_mutex_unlock(&comp_done_lock);
> -
> -            qemu_mutex_lock(&param->mutex);
> -        } else {
> -            qemu_cond_wait(&param->cond, &param->mutex);
> -        }
> -    }
> -    qemu_mutex_unlock(&param->mutex);
> -
> -    return NULL;
> -}
> -
> -static void compress_threads_save_cleanup(void)
> -{
> -    int i, thread_count;
> -
> -    if (!migrate_use_compression() || !comp_param) {
> -        return;
> -    }
> -
> -    thread_count = migrate_compress_threads();
> -    for (i = 0; i < thread_count; i++) {
> -        /*
> -         * we use it as a indicator which shows if the thread is
> -         * properly init'd or not
> -         */
> -        if (!comp_param[i].file) {
> -            break;
> -        }
> -
> -        qemu_mutex_lock(&comp_param[i].mutex);
> -        comp_param[i].quit = true;
> -        qemu_cond_signal(&comp_param[i].cond);
> -        qemu_mutex_unlock(&comp_param[i].mutex);
> -
> -        qemu_thread_join(compress_threads + i);
> -        qemu_mutex_destroy(&comp_param[i].mutex);
> -        qemu_cond_destroy(&comp_param[i].cond);
> -        deflateEnd(&comp_param[i].stream);
> -        g_free(comp_param[i].originbuf);
> -        qemu_fclose(comp_param[i].file);
> -        comp_param[i].file = NULL;
> -    }
> -    qemu_mutex_destroy(&comp_done_lock);
> -    qemu_cond_destroy(&comp_done_cond);
> -    g_free(compress_threads);
> -    g_free(comp_param);
> -    compress_threads = NULL;
> -    comp_param = NULL;
> -}
> -
> -static int compress_threads_save_setup(void)
> -{
> -    int i, thread_count;
> -
> -    if (!migrate_use_compression()) {
> -        return 0;
> -    }
> -    thread_count = migrate_compress_threads();
> -    compress_threads = g_new0(QemuThread, thread_count);
> -    comp_param = g_new0(CompressParam, thread_count);
> -    qemu_cond_init(&comp_done_cond);
> -    qemu_mutex_init(&comp_done_lock);
> -    for (i = 0; i < thread_count; i++) {
> -        comp_param[i].originbuf = g_try_malloc(TARGET_PAGE_SIZE);
> -        if (!comp_param[i].originbuf) {
> -            goto exit;
> -        }
> -
> -        if (deflateInit(&comp_param[i].stream,
> -                        migrate_compress_level()) != Z_OK) {
> -            g_free(comp_param[i].originbuf);
> -            goto exit;
> -        }
> -
> -        /* comp_param[i].file is just used as a dummy buffer to save data,
> -         * set its ops to empty.
> -         */
> -        comp_param[i].file = qemu_fopen_ops(NULL, &empty_ops);
> -        comp_param[i].done = true;
> -        comp_param[i].quit = false;
> -        qemu_mutex_init(&comp_param[i].mutex);
> -        qemu_cond_init(&comp_param[i].cond);
> -        qemu_thread_create(compress_threads + i, "compress",
> -                           do_data_compress, comp_param + i,
> -                           QEMU_THREAD_JOINABLE);
> -    }
> -    return 0;
> -
> -exit:
> -    compress_threads_save_cleanup();
> -    return -1;
> -}
> -
>  /**
>   * save_page_header: write page header to wire
>   *
> @@ -853,13 +678,12 @@ uint64_t ram_pagesize_summary(void)
>  uint64_t ram_get_total_transferred_pages(void)
>  {
>      return  ram_counters.normal + ram_counters.duplicate +
> -                compression_counters.pages + xbzrle_counters.pages;
> +            xbzrle_counters.pages;
>  }
>  
>  static void migration_update_rates(RAMState *rs, int64_t end_time)
>  {
>      uint64_t page_count = rs->target_page_count - rs->target_page_count_prev;
> -    double compressed_size;
>  
>      /* calculate period counters */
>      ram_counters.dirty_pages_rate = rs->num_dirty_pages_period * 1000
> @@ -874,26 +698,6 @@ static void migration_update_rates(RAMState *rs, int64_t 
> end_time)
>              rs->xbzrle_cache_miss_prev) / page_count;
>          rs->xbzrle_cache_miss_prev = xbzrle_counters.cache_miss;
>      }
> -
> -    if (migrate_use_compression()) {
> -        compression_counters.busy_rate = (double)(compression_counters.busy -
> -            rs->compress_thread_busy_prev) / page_count;
> -        rs->compress_thread_busy_prev = compression_counters.busy;
> -
> -        compressed_size = compression_counters.compressed_size -
> -                          rs->compressed_size_prev;
> -        if (compressed_size) {
> -            double uncompressed_size = (compression_counters.pages -
> -                                    rs->compress_pages_prev) * 
> TARGET_PAGE_SIZE;
> -
> -            /* Compression-Ratio = Uncompressed-size / Compressed-size */
> -            compression_counters.compression_rate =
> -                                        uncompressed_size / compressed_size;
> -
> -            rs->compress_pages_prev = compression_counters.pages;
> -            rs->compressed_size_prev = compression_counters.compressed_size;
> -        }
> -    }
>  }
>  
>  static void migration_bitmap_sync(RAMState *rs)
> @@ -1161,131 +965,6 @@ static int ram_save_multifd_page(RAMState *rs, 
> RAMBlock *block,
>      return 1;
>  }
>  
> -static bool do_compress_ram_page(QEMUFile *f, z_stream *stream, RAMBlock 
> *block,
> -                                 ram_addr_t offset, uint8_t *source_buf)
> -{
> -    RAMState *rs = ram_state;
> -    uint8_t *p = block->host + (offset & TARGET_PAGE_MASK);
> -    bool zero_page = false;
> -    int ret;
> -
> -    if (save_zero_page_to_file(rs, f, block, offset)) {
> -        zero_page = true;
> -        goto exit;
> -    }
> -
> -    save_page_header(rs, f, block, offset | RAM_SAVE_FLAG_COMPRESS_PAGE);
> -
> -    /*
> -     * copy it to a internal buffer to avoid it being modified by VM
> -     * so that we can catch up the error during compression and
> -     * decompression
> -     */
> -    memcpy(source_buf, p, TARGET_PAGE_SIZE);
> -    ret = qemu_put_compression_data(f, stream, source_buf, TARGET_PAGE_SIZE);
> -    if (ret < 0) {
> -        qemu_file_set_error(migrate_get_current()->to_dst_file, ret);
> -        error_report("compressed data failed!");
> -        return false;
> -    }
> -
> -exit:
> -    ram_release_pages(block->idstr, offset & TARGET_PAGE_MASK, 1);
> -    return zero_page;
> -}
> -
> -static void
> -update_compress_thread_counts(const CompressParam *param, int bytes_xmit)
> -{
> -    ram_counters.transferred += bytes_xmit;
> -
> -    if (param->zero_page) {
> -        ram_counters.duplicate++;
> -        return;
> -    }
> -
> -    /* 8 means a header with RAM_SAVE_FLAG_CONTINUE. */
> -    compression_counters.compressed_size += bytes_xmit - 8;
> -    compression_counters.pages++;
> -}
> -
> -static bool save_page_use_compression(RAMState *rs);
> -
> -static void flush_compressed_data(RAMState *rs)
> -{
> -    int idx, len, thread_count;
> -
> -    if (!save_page_use_compression(rs)) {
> -        return;
> -    }
> -    thread_count = migrate_compress_threads();
> -
> -    qemu_mutex_lock(&comp_done_lock);
> -    for (idx = 0; idx < thread_count; idx++) {
> -        while (!comp_param[idx].done) {
> -            qemu_cond_wait(&comp_done_cond, &comp_done_lock);
> -        }
> -    }
> -    qemu_mutex_unlock(&comp_done_lock);
> -
> -    for (idx = 0; idx < thread_count; idx++) {
> -        qemu_mutex_lock(&comp_param[idx].mutex);
> -        if (!comp_param[idx].quit) {
> -            len = qemu_put_qemu_file(rs->f, comp_param[idx].file);
> -            /*
> -             * it's safe to fetch zero_page without holding comp_done_lock
> -             * as there is no further request submitted to the thread,
> -             * i.e, the thread should be waiting for a request at this point.
> -             */
> -            update_compress_thread_counts(&comp_param[idx], len);
> -        }
> -        qemu_mutex_unlock(&comp_param[idx].mutex);
> -    }
> -}
> -
> -static inline void set_compress_params(CompressParam *param, RAMBlock *block,
> -                                       ram_addr_t offset)
> -{
> -    param->block = block;
> -    param->offset = offset;
> -}
> -
> -static int compress_page_with_multi_thread(RAMState *rs, RAMBlock *block,
> -                                           ram_addr_t offset)
> -{
> -    int idx, thread_count, bytes_xmit = -1, pages = -1;
> -    bool wait = migrate_compress_wait_thread();
> -
> -    thread_count = migrate_compress_threads();
> -    qemu_mutex_lock(&comp_done_lock);
> -retry:
> -    for (idx = 0; idx < thread_count; idx++) {
> -        if (comp_param[idx].done) {
> -            comp_param[idx].done = false;
> -            bytes_xmit = qemu_put_qemu_file(rs->f, comp_param[idx].file);
> -            qemu_mutex_lock(&comp_param[idx].mutex);
> -            set_compress_params(&comp_param[idx], block, offset);
> -            qemu_cond_signal(&comp_param[idx].cond);
> -            qemu_mutex_unlock(&comp_param[idx].mutex);
> -            pages = 1;
> -            update_compress_thread_counts(&comp_param[idx], bytes_xmit);
> -            break;
> -        }
> -    }
> -
> -    /*
> -     * wait for the free thread if the user specifies 'compress-wait-thread',
> -     * otherwise we will post the page out in the main thread as normal page.
> -     */
> -    if (pages < 0 && wait) {
> -        qemu_cond_wait(&comp_done_cond, &comp_done_lock);
> -        goto retry;
> -    }
> -    qemu_mutex_unlock(&comp_done_lock);
> -
> -    return pages;
> -}
> -
>  /**
>   * find_dirty_block: find the next dirty page and update any state
>   * associated with the search process.
> @@ -1314,17 +993,6 @@ static bool find_dirty_block(RAMState *rs, 
> PageSearchStatus *pss, bool *again)
>          pss->page = 0;
>          pss->block = QLIST_NEXT_RCU(pss->block, next);
>          if (!pss->block) {
> -            /*
> -             * If memory migration starts over, we will meet a dirtied page
> -             * which may still exists in compression threads's ring, so we
> -             * should flush the compressed data to make sure the new page
> -             * is not overwritten by the old one in the destination.
> -             *
> -             * Also If xbzrle is on, stop using the data compression at this
> -             * point. In theory, xbzrle can do better than compression.
> -             */
> -            flush_compressed_data(rs);
> -
>              /* Hit the end of the list */
>              pss->block = QLIST_FIRST_RCU(&ram_list.blocks);
>              /* Flag that we've looped */
> @@ -1535,58 +1203,6 @@ int ram_save_queue_pages(const char *rbname, 
> ram_addr_t start, ram_addr_t len)
>      return 0;
>  }
>  
> -static bool save_page_use_compression(RAMState *rs)
> -{
> -    if (!migrate_use_compression()) {
> -        return false;
> -    }
> -
> -    /*
> -     * If xbzrle is on, stop using the data compression after first
> -     * round of migration even if compression is enabled. In theory,
> -     * xbzrle can do better than compression.
> -     */
> -    if (rs->ram_bulk_stage || !migrate_use_xbzrle()) {
> -        return true;
> -    }
> -
> -    return false;
> -}
> -
> -/*
> - * try to compress the page before posting it out, return true if the page
> - * has been properly handled by compression, otherwise needs other
> - * paths to handle it
> - */
> -static bool save_compress_page(RAMState *rs, RAMBlock *block, ram_addr_t 
> offset)
> -{
> -    if (!save_page_use_compression(rs)) {
> -        return false;
> -    }
> -
> -    /*
> -     * When starting the process of a new block, the first page of
> -     * the block should be sent out before other pages in the same
> -     * block, and all the pages in last block should have been sent
> -     * out, keeping this order is important, because the 'cont' flag
> -     * is used to avoid resending the block name.
> -     *
> -     * We post the fist page as normal page as compression will take
> -     * much CPU resource.
> -     */
> -    if (block != rs->last_sent_block) {
> -        flush_compressed_data(rs);
> -        return false;
> -    }
> -
> -    if (compress_page_with_multi_thread(rs, block, offset) > 0) {
> -        return true;
> -    }
> -
> -    compression_counters.busy++;
> -    return false;
> -}
> -
>  /**
>   * ram_save_target_page: save one target page
>   *
> @@ -1607,16 +1223,12 @@ static int ram_save_target_page(RAMState *rs, 
> PageSearchStatus *pss,
>          return res;
>      }
>  
> -    if (save_compress_page(rs, block, offset)) {
> -        return 1;
> -    }
> -
>      res = save_zero_page(rs, block, offset);
>      if (res > 0) {
>          /* Must let xbzrle know, otherwise a previous (now 0'd) cached
>           * page would be stale
>           */
> -        if (!save_page_use_compression(rs)) {
> +        if (migrate_use_xbzrle()) {
>              XBZRLE_cache_lock();
>              xbzrle_cache_zero_page(rs, block->offset + offset);
>              XBZRLE_cache_unlock();
> @@ -1631,7 +1243,7 @@ static int ram_save_target_page(RAMState *rs, 
> PageSearchStatus *pss,
>       *    before sending the compressed page
>       * 2. In postcopy as one whole host page should be placed
>       */
> -    if (!save_page_use_compression(rs) && migrate_use_multifd()
> +    if (!migrate_use_xbzrle() && migrate_use_multifd()
>          && !migration_in_postcopy()) {
>          return ram_save_multifd_page(rs, block, offset);
>      }
> @@ -1841,7 +1453,6 @@ static void ram_save_cleanup(void *opaque)
>      }
>  
>      xbzrle_cleanup();
> -    compress_threads_save_cleanup();
>      ram_state_cleanup(rsp);
>  }
>  
> @@ -2418,14 +2029,9 @@ static int ram_save_setup(QEMUFile *f, void *opaque)
>      RAMState **rsp = opaque;
>      RAMBlock *block;
>  
> -    if (compress_threads_save_setup()) {
> -        return -1;
> -    }
> -
>      /* migration has already setup the bitmap, reuse it. */
>      if (!migration_in_colo_state()) {
>          if (ram_init_all(rsp) != 0) {
> -            compress_threads_save_cleanup();
>              return -1;
>          }
>      }
> @@ -2516,14 +2122,6 @@ static int ram_save_iterate(QEMUFile *f, void *opaque)
>  
>              rs->target_page_count += pages;
>  
> -            /*
> -             * During postcopy, it is necessary to make sure one whole host
> -             * page is sent in one chunk.
> -             */
> -            if (migrate_postcopy_ram()) {
> -                flush_compressed_data(rs);
> -            }
> -
>              /*
>               * we want to check in the 1st loop, just in case it was the 1st
>               * time and we had to sync the dirty bitmap.
> @@ -2605,7 +2203,6 @@ static int ram_save_complete(QEMUFile *f, void *opaque)
>              }
>          }
>  
> -        flush_compressed_data(rs);
>          ram_control_after_iterate(f, RAM_CONTROL_FINISH);
>      }
>  
> @@ -2773,193 +2370,6 @@ void ram_handle_compressed(void *host, uint8_t ch, 
> uint64_t size)
>      }
>  }
>  
> -/* return the size after decompression, or negative value on error */
> -static int
> -qemu_uncompress_data(z_stream *stream, uint8_t *dest, size_t dest_len,
> -                     const uint8_t *source, size_t source_len)
> -{
> -    int err;
> -
> -    err = inflateReset(stream);
> -    if (err != Z_OK) {
> -        return -1;
> -    }
> -
> -    stream->avail_in = source_len;
> -    stream->next_in = (uint8_t *)source;
> -    stream->avail_out = dest_len;
> -    stream->next_out = dest;
> -
> -    err = inflate(stream, Z_NO_FLUSH);
> -    if (err != Z_STREAM_END) {
> -        return -1;
> -    }
> -
> -    return stream->total_out;
> -}
> -
> -static void *do_data_decompress(void *opaque)
> -{
> -    DecompressParam *param = opaque;
> -    unsigned long pagesize;
> -    uint8_t *des;
> -    int len, ret;
> -
> -    qemu_mutex_lock(&param->mutex);
> -    while (!param->quit) {
> -        if (param->des) {
> -            des = param->des;
> -            len = param->len;
> -            param->des = 0;
> -            qemu_mutex_unlock(&param->mutex);
> -
> -            pagesize = TARGET_PAGE_SIZE;
> -
> -            ret = qemu_uncompress_data(&param->stream, des, pagesize,
> -                                       param->compbuf, len);
> -            if (ret < 0 && migrate_get_current()->decompress_error_check) {
> -                error_report("decompress data failed");
> -                qemu_file_set_error(decomp_file, ret);
> -            }
> -
> -            qemu_mutex_lock(&decomp_done_lock);
> -            param->done = true;
> -            qemu_cond_signal(&decomp_done_cond);
> -            qemu_mutex_unlock(&decomp_done_lock);
> -
> -            qemu_mutex_lock(&param->mutex);
> -        } else {
> -            qemu_cond_wait(&param->cond, &param->mutex);
> -        }
> -    }
> -    qemu_mutex_unlock(&param->mutex);
> -
> -    return NULL;
> -}
> -
> -static int wait_for_decompress_done(void)
> -{
> -    int idx, thread_count;
> -
> -    if (!migrate_use_compression()) {
> -        return 0;
> -    }
> -
> -    thread_count = migrate_decompress_threads();
> -    qemu_mutex_lock(&decomp_done_lock);
> -    for (idx = 0; idx < thread_count; idx++) {
> -        while (!decomp_param[idx].done) {
> -            qemu_cond_wait(&decomp_done_cond, &decomp_done_lock);
> -        }
> -    }
> -    qemu_mutex_unlock(&decomp_done_lock);
> -    return qemu_file_get_error(decomp_file);
> -}
> -
> -static void compress_threads_load_cleanup(void)
> -{
> -    int i, thread_count;
> -
> -    if (!migrate_use_compression()) {
> -        return;
> -    }
> -    thread_count = migrate_decompress_threads();
> -    for (i = 0; i < thread_count; i++) {
> -        /*
> -         * we use it as a indicator which shows if the thread is
> -         * properly init'd or not
> -         */
> -        if (!decomp_param[i].compbuf) {
> -            break;
> -        }
> -
> -        qemu_mutex_lock(&decomp_param[i].mutex);
> -        decomp_param[i].quit = true;
> -        qemu_cond_signal(&decomp_param[i].cond);
> -        qemu_mutex_unlock(&decomp_param[i].mutex);
> -    }
> -    for (i = 0; i < thread_count; i++) {
> -        if (!decomp_param[i].compbuf) {
> -            break;
> -        }
> -
> -        qemu_thread_join(decompress_threads + i);
> -        qemu_mutex_destroy(&decomp_param[i].mutex);
> -        qemu_cond_destroy(&decomp_param[i].cond);
> -        inflateEnd(&decomp_param[i].stream);
> -        g_free(decomp_param[i].compbuf);
> -        decomp_param[i].compbuf = NULL;
> -    }
> -    g_free(decompress_threads);
> -    g_free(decomp_param);
> -    decompress_threads = NULL;
> -    decomp_param = NULL;
> -    decomp_file = NULL;
> -}
> -
> -static int compress_threads_load_setup(QEMUFile *f)
> -{
> -    int i, thread_count;
> -
> -    if (!migrate_use_compression()) {
> -        return 0;
> -    }
> -
> -    thread_count = migrate_decompress_threads();
> -    decompress_threads = g_new0(QemuThread, thread_count);
> -    decomp_param = g_new0(DecompressParam, thread_count);
> -    qemu_mutex_init(&decomp_done_lock);
> -    qemu_cond_init(&decomp_done_cond);
> -    decomp_file = f;
> -    for (i = 0; i < thread_count; i++) {
> -        if (inflateInit(&decomp_param[i].stream) != Z_OK) {
> -            goto exit;
> -        }
> -
> -        decomp_param[i].compbuf = g_malloc0(compressBound(TARGET_PAGE_SIZE));
> -        qemu_mutex_init(&decomp_param[i].mutex);
> -        qemu_cond_init(&decomp_param[i].cond);
> -        decomp_param[i].done = true;
> -        decomp_param[i].quit = false;
> -        qemu_thread_create(decompress_threads + i, "decompress",
> -                           do_data_decompress, decomp_param + i,
> -                           QEMU_THREAD_JOINABLE);
> -    }
> -    return 0;
> -exit:
> -    compress_threads_load_cleanup();
> -    return -1;
> -}
> -
> -static void decompress_data_with_multi_threads(QEMUFile *f,
> -                                               void *host, int len)
> -{
> -    int idx, thread_count;
> -
> -    thread_count = migrate_decompress_threads();
> -    qemu_mutex_lock(&decomp_done_lock);
> -    while (true) {
> -        for (idx = 0; idx < thread_count; idx++) {
> -            if (decomp_param[idx].done) {
> -                decomp_param[idx].done = false;
> -                qemu_mutex_lock(&decomp_param[idx].mutex);
> -                qemu_get_buffer(f, decomp_param[idx].compbuf, len);
> -                decomp_param[idx].des = host;
> -                decomp_param[idx].len = len;
> -                qemu_cond_signal(&decomp_param[idx].cond);
> -                qemu_mutex_unlock(&decomp_param[idx].mutex);
> -                break;
> -            }
> -        }
> -        if (idx < thread_count) {
> -            break;
> -        } else {
> -            qemu_cond_wait(&decomp_done_cond, &decomp_done_lock);
> -        }
> -    }
> -    qemu_mutex_unlock(&decomp_done_lock);
> -}
> -
>  /*
>   * colo cache: this is for secondary VM, we cache the whole
>   * memory of the secondary VM, it is need to hold the global lock
> @@ -3047,10 +2457,6 @@ void colo_release_ram_cache(void)
>   */
>  static int ram_load_setup(QEMUFile *f, void *opaque)
>  {
> -    if (compress_threads_load_setup(f)) {
> -        return -1;
> -    }
> -
>      xbzrle_load_setup();
>      ramblock_recv_map_init();
>  
> @@ -3066,7 +2472,6 @@ static int ram_load_cleanup(void *opaque)
>      }
>  
>      xbzrle_load_cleanup();
> -    compress_threads_load_cleanup();
>  
>      RAMBLOCK_FOREACH_NOT_IGNORED(rb) {
>          g_free(rb->receivedmap);
> @@ -3121,7 +2526,6 @@ static int ram_load_postcopy(QEMUFile *f)
>          void *place_source = NULL;
>          RAMBlock *block = NULL;
>          uint8_t ch;
> -        int len;
>  
>          addr = qemu_get_be64(f);
>  
> @@ -3139,8 +2543,7 @@ static int ram_load_postcopy(QEMUFile *f)
>  
>          trace_ram_load_postcopy_loop((uint64_t)addr, flags);
>          place_needed = false;
> -        if (flags & (RAM_SAVE_FLAG_ZERO | RAM_SAVE_FLAG_PAGE |
> -                     RAM_SAVE_FLAG_COMPRESS_PAGE)) {
> +        if (flags & (RAM_SAVE_FLAG_ZERO | RAM_SAVE_FLAG_PAGE)) {
>              block = ram_block_from_stream(f, flags);
>  
>              host = host_from_ram_block_offset(block, addr);
> @@ -3223,17 +2626,6 @@ static int ram_load_postcopy(QEMUFile *f)
>                                           TARGET_PAGE_SIZE);
>              }
>              break;
> -        case RAM_SAVE_FLAG_COMPRESS_PAGE:
> -            all_zero = false;
> -            len = qemu_get_be32(f);
> -            if (len < 0 || len > compressBound(TARGET_PAGE_SIZE)) {
> -                error_report("Invalid compressed data length: %d", len);
> -                ret = -EINVAL;
> -                break;
> -            }
> -            decompress_data_with_multi_threads(f, page_buffer, len);
> -            break;
> -
>          case RAM_SAVE_FLAG_EOS:
>              /* normal exit */
>              multifd_recv_sync_main();
> @@ -3245,11 +2637,6 @@ static int ram_load_postcopy(QEMUFile *f)
>              break;
>          }
>  
> -        /* Got the whole host page, wait for decompress before placing. */
> -        if (place_needed) {
> -            ret |= wait_for_decompress_done();
> -        }
> -
>          /* Detect for any possible file errors */
>          if (!ret && qemu_file_get_error(f)) {
>              ret = qemu_file_get_error(f);
> @@ -3339,12 +2726,9 @@ static void colo_flush_ram_cache(void)
>   */
>  static int ram_load_precopy(QEMUFile *f)
>  {
> -    int flags = 0, ret = 0, invalid_flags = 0, len = 0, i = 0;
> +    int flags = 0, ret = 0, len = 0, i = 0;
>      /* ADVISE is earlier, it shows the source has the postcopy capability on 
> */
>      bool postcopy_advised = postcopy_is_advised();
> -    if (!migrate_use_compression()) {
> -        invalid_flags |= RAM_SAVE_FLAG_COMPRESS_PAGE;
> -    }
>  
>      while (!ret && !(flags & RAM_SAVE_FLAG_EOS)) {
>          ram_addr_t addr, total_ram_bytes;
> @@ -3366,17 +2750,8 @@ static int ram_load_precopy(QEMUFile *f)
>          flags = addr & ~TARGET_PAGE_MASK;
>          addr &= TARGET_PAGE_MASK;
>  
> -        if (flags & invalid_flags) {
> -            if (flags & invalid_flags & RAM_SAVE_FLAG_COMPRESS_PAGE) {
> -                error_report("Received an unexpected compressed page");
> -            }
> -
> -            ret = -EINVAL;
> -            break;
> -        }
> -
>          if (flags & (RAM_SAVE_FLAG_ZERO | RAM_SAVE_FLAG_PAGE |
> -                     RAM_SAVE_FLAG_COMPRESS_PAGE | RAM_SAVE_FLAG_XBZRLE)) {
> +                     RAM_SAVE_FLAG_XBZRLE)) {
>              RAMBlock *block = ram_block_from_stream(f, flags);
>  
>              /*
> @@ -3472,16 +2847,6 @@ static int ram_load_precopy(QEMUFile *f)
>              qemu_get_buffer(f, host, TARGET_PAGE_SIZE);
>              break;
>  
> -        case RAM_SAVE_FLAG_COMPRESS_PAGE:
> -            len = qemu_get_be32(f);
> -            if (len < 0 || len > compressBound(TARGET_PAGE_SIZE)) {
> -                error_report("Invalid compressed data length: %d", len);
> -                ret = -EINVAL;
> -                break;
> -            }
> -            decompress_data_with_multi_threads(f, host, len);
> -            break;
> -
>          case RAM_SAVE_FLAG_XBZRLE:
>              if (load_xbzrle(f, addr, host) < 0) {
>                  error_report("Failed to decompress XBZRLE page at "
> @@ -3508,7 +2873,6 @@ static int ram_load_precopy(QEMUFile *f)
>          }
>      }
>  
> -    ret |= wait_for_decompress_done();
>      return ret;
>  }
>  
> diff --git a/migration/ram.h b/migration/ram.h
> index a553d40751..638e74e499 100644
> --- a/migration/ram.h
> +++ b/migration/ram.h
> @@ -35,7 +35,6 @@
>  
>  extern MigrationStats ram_counters;
>  extern XBZRLECacheStats xbzrle_counters;
> -extern CompressionStats compression_counters;
>  
>  int xbzrle_cache_resize(int64_t new_size, Error **errp);
>  uint64_t ram_bytes_remaining(void);
> diff --git a/monitor/hmp-cmds.c b/monitor/hmp-cmds.c
> index ba75a609a7..185572f1ed 100644
> --- a/monitor/hmp-cmds.c
> +++ b/monitor/hmp-cmds.c
> @@ -309,19 +309,6 @@ void hmp_info_migrate(Monitor *mon, const QDict *qdict)
>                         info->xbzrle_cache->overflow);
>      }
>  
> -    if (info->has_compression) {
> -        monitor_printf(mon, "compression pages: %" PRIu64 " pages\n",
> -                       info->compression->pages);
> -        monitor_printf(mon, "compression busy: %" PRIu64 "\n",
> -                       info->compression->busy);
> -        monitor_printf(mon, "compression busy rate: %0.2f\n",
> -                       info->compression->busy_rate);
> -        monitor_printf(mon, "compressed size: %" PRIu64 "\n",
> -                       info->compression->compressed_size);
> -        monitor_printf(mon, "compression rate: %0.2f\n",
> -                       info->compression->compression_rate);
> -    }
> -
>      if (info->has_cpu_throttle_percentage) {
>          monitor_printf(mon, "cpu throttle percentage: %" PRIu64 "\n",
>                         info->cpu_throttle_percentage);
> @@ -397,18 +384,6 @@ void hmp_info_migrate_parameters(Monitor *mon, const 
> QDict *qdict)
>          monitor_printf(mon, "%s: %u\n",
>              MigrationParameter_str(MIGRATION_PARAMETER_COMPRESS_LEVEL),
>              params->compress_level);
> -        assert(params->has_compress_threads);
> -        monitor_printf(mon, "%s: %u\n",
> -            MigrationParameter_str(MIGRATION_PARAMETER_COMPRESS_THREADS),
> -            params->compress_threads);
> -        assert(params->has_compress_wait_thread);
> -        monitor_printf(mon, "%s: %s\n",
> -            MigrationParameter_str(MIGRATION_PARAMETER_COMPRESS_WAIT_THREAD),
> -            params->compress_wait_thread ? "on" : "off");
> -        assert(params->has_decompress_threads);
> -        monitor_printf(mon, "%s: %u\n",
> -            MigrationParameter_str(MIGRATION_PARAMETER_DECOMPRESS_THREADS),
> -            params->decompress_threads);
>          assert(params->has_cpu_throttle_initial);
>          monitor_printf(mon, "%s: %u\n",
>              MigrationParameter_str(MIGRATION_PARAMETER_CPU_THROTTLE_INITIAL),
> @@ -1756,18 +1731,6 @@ void hmp_migrate_set_parameter(Monitor *mon, const 
> QDict *qdict)
>          p->has_compress_level = true;
>          visit_type_int(v, param, &p->compress_level, &err);
>          break;
> -    case MIGRATION_PARAMETER_COMPRESS_THREADS:
> -        p->has_compress_threads = true;
> -        visit_type_int(v, param, &p->compress_threads, &err);
> -        break;
> -    case MIGRATION_PARAMETER_COMPRESS_WAIT_THREAD:
> -        p->has_compress_wait_thread = true;
> -        visit_type_bool(v, param, &p->compress_wait_thread, &err);
> -        break;
> -    case MIGRATION_PARAMETER_DECOMPRESS_THREADS:
> -        p->has_decompress_threads = true;
> -        visit_type_int(v, param, &p->decompress_threads, &err);
> -        break;
>      case MIGRATION_PARAMETER_CPU_THROTTLE_INITIAL:
>          p->has_cpu_throttle_initial = true;
>          visit_type_int(v, param, &p->cpu_throttle_initial, &err);
> diff --git a/qapi/migration.json b/qapi/migration.json
> index 65db85970e..5477d4d20b 100644
> --- a/qapi/migration.json
> +++ b/qapi/migration.json
> @@ -79,27 +79,6 @@
>             'cache-miss': 'int', 'cache-miss-rate': 'number',
>             'overflow': 'int' } }
>  
> -##
> -# @CompressionStats:
> -#
> -# Detailed migration compression statistics
> -#
> -# @pages: amount of pages compressed and transferred to the target VM
> -#
> -# @busy: count of times that no free thread was available to compress data
> -#
> -# @busy-rate: rate of thread busy
> -#
> -# @compressed-size: amount of bytes after compression
> -#
> -# @compression-rate: rate of compressed size
> -#
> -# Since: 3.1
> -##
> -{ 'struct': 'CompressionStats',
> -  'data': {'pages': 'int', 'busy': 'int', 'busy-rate': 'number',
> -        'compressed-size': 'int', 'compression-rate': 'number' } }
> -
>  ##
>  # @MigrationStatus:
>  #
> @@ -200,9 +179,6 @@
>  #           only present when the postcopy-blocktime migration capability
>  #           is enabled. (Since 3.0)
>  #
> -# @compression: migration compression statistics, only returned if 
> compression
> -#           feature is on and status is 'active' or 'completed' (Since 3.1)
> -#
>  # @socket-address: Only used for tcp, to know what the real port is (Since 
> 4.0)
>  #
>  # Since: 0.14.0
> @@ -219,7 +195,6 @@
>             '*error-desc': 'str',
>             '*postcopy-blocktime' : 'uint32',
>             '*postcopy-vcpu-blocktime': ['uint32'],
> -           '*compression': 'CompressionStats',
>             '*socket-address': ['SocketAddress'] } }
>  
>  ##
> @@ -364,14 +339,6 @@
>  #          to enable the capability on the source VM. The feature is 
> disabled by
>  #          default. (since 1.6)
>  #
> -# @compress: Use multiple compression threads to accelerate live migration.
> -#          This feature can help to reduce the migration traffic, by sending
> -#          compressed pages. Please note that if compress and xbzrle are both
> -#          on, compress only takes effect in the ram bulk stage, after that,
> -#          it will be disabled and only xbzrle takes effect, this can help to
> -#          minimize migration traffic. The feature is disabled by default.
> -#          (since 2.4 )
> -#
>  # @events: generate events for each migration state change
>  #          (since 2.4 )
>  #
> @@ -425,7 +392,7 @@
>  ##
>  { 'enum': 'MigrationCapability',
>    'data': ['xbzrle', 'rdma-pin-all', 'auto-converge', 'zero-blocks',
> -           'compress', 'events', 'postcopy-ram', 'x-colo', 'release-ram',
> +           'events', 'postcopy-ram', 'x-colo', 'release-ram',
>             'block', 'return-path', 'pause-before-switchover', 'multifd',
>             'dirty-bitmaps', 'postcopy-blocktime', 'late-block-activate',
>             'x-ignore-shared', 'validate-uuid' ] }
> @@ -479,7 +446,6 @@
>  #       {"state": false, "capability": "rdma-pin-all"},
>  #       {"state": false, "capability": "auto-converge"},
>  #       {"state": false, "capability": "zero-blocks"},
> -#       {"state": false, "capability": "compress"},
>  #       {"state": true, "capability": "events"},
>  #       {"state": false, "capability": "postcopy-ram"},
>  #       {"state": false, "capability": "x-colo"}
> @@ -523,20 +489,6 @@
>  #          no compression, 1 means the best compression speed, and 9 means 
> best
>  #          compression ratio which will consume more CPU.
>  #
> -# @compress-threads: Set compression thread count to be used in live 
> migration,
> -#          the compression thread count is an integer between 1 and 255.
> -#
> -# @compress-wait-thread: Controls behavior when all compression threads are
> -#                        currently busy. If true (default), wait for a free
> -#                        compression thread to become available; otherwise,
> -#                        send the page uncompressed. (Since 3.1)
> -#
> -# @decompress-threads: Set decompression thread count to be used in live
> -#          migration, the decompression thread count is an integer between 1
> -#          and 255. Usually, decompression is at least 4 times as fast as
> -#          compression, so set the decompress-threads to the number about 1/4
> -#          of compress-threads is adequate.
> -#
>  # @cpu-throttle-initial: Initial percentage of time guest cpus are throttled
>  #                        when migration auto-converge is activated. The
>  #                        default value is 20. (Since 2.7)
> @@ -607,8 +559,7 @@
>  { 'enum': 'MigrationParameter',
>    'data': ['announce-initial', 'announce-max',
>             'announce-rounds', 'announce-step',
> -           'compress-level', 'compress-threads', 'decompress-threads',
> -           'compress-wait-thread',
> +           'compress-level',
>             'cpu-throttle-initial', 'cpu-throttle-increment',
>             'tls-creds', 'tls-hostname', 'tls-authz', 'max-bandwidth',
>             'downtime-limit', 'x-checkpoint-delay', 'block-incremental',
> @@ -633,15 +584,6 @@
>  #
>  # @compress-level: compression level
>  #
> -# @compress-threads: compression thread count
> -#
> -# @compress-wait-thread: Controls behavior when all compression threads are
> -#                        currently busy. If true (default), wait for a free
> -#                        compression thread to become available; otherwise,
> -#                        send the page uncompressed. (Since 3.1)
> -#
> -# @decompress-threads: decompression thread count
> -#
>  # @cpu-throttle-initial: Initial percentage of time guest cpus are
>  #                        throttled when migration auto-converge is activated.
>  #                        The default value is 20. (Since 2.7)
> @@ -717,9 +659,6 @@
>              '*announce-rounds': 'size',
>              '*announce-step': 'size',
>              '*compress-level': 'int',
> -            '*compress-threads': 'int',
> -            '*compress-wait-thread': 'bool',
> -            '*decompress-threads': 'int',
>              '*cpu-throttle-initial': 'int',
>              '*cpu-throttle-increment': 'int',
>              '*tls-creds': 'StrOrNull',
> @@ -770,15 +709,6 @@
>  #
>  # @compress-level: compression level
>  #
> -# @compress-threads: compression thread count
> -#
> -# @compress-wait-thread: Controls behavior when all compression threads are
> -#                        currently busy. If true (default), wait for a free
> -#                        compression thread to become available; otherwise,
> -#                        send the page uncompressed. (Since 3.1)
> -#
> -# @decompress-threads: decompression thread count
> -#
>  # @cpu-throttle-initial: Initial percentage of time guest cpus are
>  #                        throttled when migration auto-converge is activated.
>  #                        (Since 2.7)
> @@ -854,9 +784,6 @@
>              '*announce-rounds': 'size',
>              '*announce-step': 'size',
>              '*compress-level': 'uint8',
> -            '*compress-threads': 'uint8',
> -            '*compress-wait-thread': 'bool',
> -            '*decompress-threads': 'uint8',
>              '*cpu-throttle-initial': 'uint8',
>              '*cpu-throttle-increment': 'uint8',
>              '*tls-creds': 'str',
> @@ -885,9 +812,7 @@
>  #
>  # -> { "execute": "query-migrate-parameters" }
>  # <- { "return": {
> -#          "decompress-threads": 2,
>  #          "cpu-throttle-increment": 10,
> -#          "compress-threads": 8,
>  #          "compress-level": 1,
>  #          "cpu-throttle-initial": 20,
>  #          "max-bandwidth": 33554432,
> -- 
> 2.24.1
> 
--
Dr. David Alan Gilbert / address@hidden / Manchester, UK




reply via email to

[Prev in Thread] Current Thread [Next in Thread]