Make ram-compress.c target independent. Signed-off-by: Lukas Straub <lukasstraub2@web.de> Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> Reviewed-by: Juan Quintela <quintela@redhat.com> Signed-off-by: Juan Quintela <quintela@redhat.com>
		
			
				
	
	
		
			486 lines
		
	
	
		
			14 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			486 lines
		
	
	
		
			14 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * QEMU System Emulator
 | 
						|
 *
 | 
						|
 * Copyright (c) 2003-2008 Fabrice Bellard
 | 
						|
 * Copyright (c) 2011-2015 Red Hat Inc
 | 
						|
 *
 | 
						|
 * Authors:
 | 
						|
 *  Juan Quintela <quintela@redhat.com>
 | 
						|
 *
 | 
						|
 * Permission is hereby granted, free of charge, to any person obtaining a copy
 | 
						|
 * of this software and associated documentation files (the "Software"), to deal
 | 
						|
 * in the Software without restriction, including without limitation the rights
 | 
						|
 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
 | 
						|
 * copies of the Software, and to permit persons to whom the Software is
 | 
						|
 * furnished to do so, subject to the following conditions:
 | 
						|
 *
 | 
						|
 * The above copyright notice and this permission notice shall be included in
 | 
						|
 * all copies or substantial portions of the Software.
 | 
						|
 *
 | 
						|
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 | 
						|
 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 | 
						|
 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
 | 
						|
 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
 | 
						|
 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
 | 
						|
 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
 | 
						|
 * THE SOFTWARE.
 | 
						|
 */
 | 
						|
 | 
						|
#include "qemu/osdep.h"
 | 
						|
#include "qemu/cutils.h"
 | 
						|
 | 
						|
#include "ram-compress.h"
 | 
						|
 | 
						|
#include "qemu/error-report.h"
 | 
						|
#include "migration.h"
 | 
						|
#include "options.h"
 | 
						|
#include "io/channel-null.h"
 | 
						|
#include "exec/target_page.h"
 | 
						|
#include "exec/ramblock.h"
 | 
						|
 | 
						|
CompressionStats compression_counters;
 | 
						|
 | 
						|
static CompressParam *comp_param;
 | 
						|
static QemuThread *compress_threads;
 | 
						|
/* comp_done_cond is used to wake up the migration thread when
 | 
						|
 * one of the compression threads has finished the compression.
 | 
						|
 * comp_done_lock is used to co-work with comp_done_cond.
 | 
						|
 */
 | 
						|
static QemuMutex comp_done_lock;
 | 
						|
static QemuCond comp_done_cond;
 | 
						|
 | 
						|
struct DecompressParam {
 | 
						|
    bool done;
 | 
						|
    bool quit;
 | 
						|
    QemuMutex mutex;
 | 
						|
    QemuCond cond;
 | 
						|
    void *des;
 | 
						|
    uint8_t *compbuf;
 | 
						|
    int len;
 | 
						|
    z_stream stream;
 | 
						|
};
 | 
						|
typedef struct DecompressParam DecompressParam;
 | 
						|
 | 
						|
static QEMUFile *decomp_file;
 | 
						|
static DecompressParam *decomp_param;
 | 
						|
static QemuThread *decompress_threads;
 | 
						|
static QemuMutex decomp_done_lock;
 | 
						|
static QemuCond decomp_done_cond;
 | 
						|
 | 
						|
static CompressResult do_compress_ram_page(QEMUFile *f, z_stream *stream,
 | 
						|
                                           RAMBlock *block, ram_addr_t offset,
 | 
						|
                                           uint8_t *source_buf);
 | 
						|
 | 
						|
static void *do_data_compress(void *opaque)
 | 
						|
{
 | 
						|
    CompressParam *param = opaque;
 | 
						|
    RAMBlock *block;
 | 
						|
    ram_addr_t offset;
 | 
						|
    CompressResult result;
 | 
						|
 | 
						|
    qemu_mutex_lock(¶m->mutex);
 | 
						|
    while (!param->quit) {
 | 
						|
        if (param->trigger) {
 | 
						|
            block = param->block;
 | 
						|
            offset = param->offset;
 | 
						|
            param->trigger = false;
 | 
						|
            qemu_mutex_unlock(¶m->mutex);
 | 
						|
 | 
						|
            result = do_compress_ram_page(param->file, ¶m->stream,
 | 
						|
                                          block, offset, param->originbuf);
 | 
						|
 | 
						|
            qemu_mutex_lock(&comp_done_lock);
 | 
						|
            param->done = true;
 | 
						|
            param->result = result;
 | 
						|
            qemu_cond_signal(&comp_done_cond);
 | 
						|
            qemu_mutex_unlock(&comp_done_lock);
 | 
						|
 | 
						|
            qemu_mutex_lock(¶m->mutex);
 | 
						|
        } else {
 | 
						|
            qemu_cond_wait(¶m->cond, ¶m->mutex);
 | 
						|
        }
 | 
						|
    }
 | 
						|
    qemu_mutex_unlock(¶m->mutex);
 | 
						|
 | 
						|
    return NULL;
 | 
						|
}
 | 
						|
 | 
						|
void compress_threads_save_cleanup(void)
 | 
						|
{
 | 
						|
    int i, thread_count;
 | 
						|
 | 
						|
    if (!migrate_compress() || !comp_param) {
 | 
						|
        return;
 | 
						|
    }
 | 
						|
 | 
						|
    thread_count = migrate_compress_threads();
 | 
						|
    for (i = 0; i < thread_count; i++) {
 | 
						|
        /*
 | 
						|
         * we use it as a indicator which shows if the thread is
 | 
						|
         * properly init'd or not
 | 
						|
         */
 | 
						|
        if (!comp_param[i].file) {
 | 
						|
            break;
 | 
						|
        }
 | 
						|
 | 
						|
        qemu_mutex_lock(&comp_param[i].mutex);
 | 
						|
        comp_param[i].quit = true;
 | 
						|
        qemu_cond_signal(&comp_param[i].cond);
 | 
						|
        qemu_mutex_unlock(&comp_param[i].mutex);
 | 
						|
 | 
						|
        qemu_thread_join(compress_threads + i);
 | 
						|
        qemu_mutex_destroy(&comp_param[i].mutex);
 | 
						|
        qemu_cond_destroy(&comp_param[i].cond);
 | 
						|
        deflateEnd(&comp_param[i].stream);
 | 
						|
        g_free(comp_param[i].originbuf);
 | 
						|
        qemu_fclose(comp_param[i].file);
 | 
						|
        comp_param[i].file = NULL;
 | 
						|
    }
 | 
						|
    qemu_mutex_destroy(&comp_done_lock);
 | 
						|
    qemu_cond_destroy(&comp_done_cond);
 | 
						|
    g_free(compress_threads);
 | 
						|
    g_free(comp_param);
 | 
						|
    compress_threads = NULL;
 | 
						|
    comp_param = NULL;
 | 
						|
}
 | 
						|
 | 
						|
int compress_threads_save_setup(void)
 | 
						|
{
 | 
						|
    int i, thread_count;
 | 
						|
 | 
						|
    if (!migrate_compress()) {
 | 
						|
        return 0;
 | 
						|
    }
 | 
						|
    thread_count = migrate_compress_threads();
 | 
						|
    compress_threads = g_new0(QemuThread, thread_count);
 | 
						|
    comp_param = g_new0(CompressParam, thread_count);
 | 
						|
    qemu_cond_init(&comp_done_cond);
 | 
						|
    qemu_mutex_init(&comp_done_lock);
 | 
						|
    for (i = 0; i < thread_count; i++) {
 | 
						|
        comp_param[i].originbuf = g_try_malloc(qemu_target_page_size());
 | 
						|
        if (!comp_param[i].originbuf) {
 | 
						|
            goto exit;
 | 
						|
        }
 | 
						|
 | 
						|
        if (deflateInit(&comp_param[i].stream,
 | 
						|
                        migrate_compress_level()) != Z_OK) {
 | 
						|
            g_free(comp_param[i].originbuf);
 | 
						|
            goto exit;
 | 
						|
        }
 | 
						|
 | 
						|
        /* comp_param[i].file is just used as a dummy buffer to save data,
 | 
						|
         * set its ops to empty.
 | 
						|
         */
 | 
						|
        comp_param[i].file = qemu_file_new_output(
 | 
						|
            QIO_CHANNEL(qio_channel_null_new()));
 | 
						|
        comp_param[i].done = true;
 | 
						|
        comp_param[i].quit = false;
 | 
						|
        qemu_mutex_init(&comp_param[i].mutex);
 | 
						|
        qemu_cond_init(&comp_param[i].cond);
 | 
						|
        qemu_thread_create(compress_threads + i, "compress",
 | 
						|
                           do_data_compress, comp_param + i,
 | 
						|
                           QEMU_THREAD_JOINABLE);
 | 
						|
    }
 | 
						|
    return 0;
 | 
						|
 | 
						|
exit:
 | 
						|
    compress_threads_save_cleanup();
 | 
						|
    return -1;
 | 
						|
}
 | 
						|
 | 
						|
static CompressResult do_compress_ram_page(QEMUFile *f, z_stream *stream,
 | 
						|
                                           RAMBlock *block, ram_addr_t offset,
 | 
						|
                                           uint8_t *source_buf)
 | 
						|
{
 | 
						|
    uint8_t *p = block->host + offset;
 | 
						|
    size_t page_size = qemu_target_page_size();
 | 
						|
    int ret;
 | 
						|
 | 
						|
    assert(qemu_file_buffer_empty(f));
 | 
						|
 | 
						|
    if (buffer_is_zero(p, page_size)) {
 | 
						|
        return RES_ZEROPAGE;
 | 
						|
    }
 | 
						|
 | 
						|
    /*
 | 
						|
     * copy it to a internal buffer to avoid it being modified by VM
 | 
						|
     * so that we can catch up the error during compression and
 | 
						|
     * decompression
 | 
						|
     */
 | 
						|
    memcpy(source_buf, p, page_size);
 | 
						|
    ret = qemu_put_compression_data(f, stream, source_buf, page_size);
 | 
						|
    if (ret < 0) {
 | 
						|
        qemu_file_set_error(migrate_get_current()->to_dst_file, ret);
 | 
						|
        error_report("compressed data failed!");
 | 
						|
        qemu_fflush(f);
 | 
						|
        return RES_NONE;
 | 
						|
    }
 | 
						|
    return RES_COMPRESS;
 | 
						|
}
 | 
						|
 | 
						|
static inline void compress_reset_result(CompressParam *param)
 | 
						|
{
 | 
						|
    param->result = RES_NONE;
 | 
						|
    param->block = NULL;
 | 
						|
    param->offset = 0;
 | 
						|
}
 | 
						|
 | 
						|
void flush_compressed_data(int (send_queued_data(CompressParam *)))
 | 
						|
{
 | 
						|
    int idx, thread_count;
 | 
						|
 | 
						|
    thread_count = migrate_compress_threads();
 | 
						|
 | 
						|
    qemu_mutex_lock(&comp_done_lock);
 | 
						|
    for (idx = 0; idx < thread_count; idx++) {
 | 
						|
        while (!comp_param[idx].done) {
 | 
						|
            qemu_cond_wait(&comp_done_cond, &comp_done_lock);
 | 
						|
        }
 | 
						|
    }
 | 
						|
    qemu_mutex_unlock(&comp_done_lock);
 | 
						|
 | 
						|
    for (idx = 0; idx < thread_count; idx++) {
 | 
						|
        qemu_mutex_lock(&comp_param[idx].mutex);
 | 
						|
        if (!comp_param[idx].quit) {
 | 
						|
            CompressParam *param = &comp_param[idx];
 | 
						|
            send_queued_data(param);
 | 
						|
            assert(qemu_file_buffer_empty(param->file));
 | 
						|
            compress_reset_result(param);
 | 
						|
        }
 | 
						|
        qemu_mutex_unlock(&comp_param[idx].mutex);
 | 
						|
    }
 | 
						|
}
 | 
						|
 | 
						|
static inline void set_compress_params(CompressParam *param, RAMBlock *block,
 | 
						|
                                       ram_addr_t offset)
 | 
						|
{
 | 
						|
    param->block = block;
 | 
						|
    param->offset = offset;
 | 
						|
    param->trigger = true;
 | 
						|
}
 | 
						|
 | 
						|
int compress_page_with_multi_thread(RAMBlock *block, ram_addr_t offset,
 | 
						|
                                int (send_queued_data(CompressParam *)))
 | 
						|
{
 | 
						|
    int idx, thread_count, pages = -1;
 | 
						|
    bool wait = migrate_compress_wait_thread();
 | 
						|
 | 
						|
    thread_count = migrate_compress_threads();
 | 
						|
    qemu_mutex_lock(&comp_done_lock);
 | 
						|
retry:
 | 
						|
    for (idx = 0; idx < thread_count; idx++) {
 | 
						|
        if (comp_param[idx].done) {
 | 
						|
            CompressParam *param = &comp_param[idx];
 | 
						|
            qemu_mutex_lock(¶m->mutex);
 | 
						|
            param->done = false;
 | 
						|
            send_queued_data(param);
 | 
						|
            assert(qemu_file_buffer_empty(param->file));
 | 
						|
            compress_reset_result(param);
 | 
						|
            set_compress_params(param, block, offset);
 | 
						|
 | 
						|
            qemu_cond_signal(¶m->cond);
 | 
						|
            qemu_mutex_unlock(¶m->mutex);
 | 
						|
            pages = 1;
 | 
						|
            break;
 | 
						|
        }
 | 
						|
    }
 | 
						|
 | 
						|
    /*
 | 
						|
     * wait for the free thread if the user specifies 'compress-wait-thread',
 | 
						|
     * otherwise we will post the page out in the main thread as normal page.
 | 
						|
     */
 | 
						|
    if (pages < 0 && wait) {
 | 
						|
        qemu_cond_wait(&comp_done_cond, &comp_done_lock);
 | 
						|
        goto retry;
 | 
						|
    }
 | 
						|
    qemu_mutex_unlock(&comp_done_lock);
 | 
						|
 | 
						|
    return pages;
 | 
						|
}
 | 
						|
 | 
						|
/* return the size after decompression, or negative value on error */
 | 
						|
static int
 | 
						|
qemu_uncompress_data(z_stream *stream, uint8_t *dest, size_t dest_len,
 | 
						|
                     const uint8_t *source, size_t source_len)
 | 
						|
{
 | 
						|
    int err;
 | 
						|
 | 
						|
    err = inflateReset(stream);
 | 
						|
    if (err != Z_OK) {
 | 
						|
        return -1;
 | 
						|
    }
 | 
						|
 | 
						|
    stream->avail_in = source_len;
 | 
						|
    stream->next_in = (uint8_t *)source;
 | 
						|
    stream->avail_out = dest_len;
 | 
						|
    stream->next_out = dest;
 | 
						|
 | 
						|
    err = inflate(stream, Z_NO_FLUSH);
 | 
						|
    if (err != Z_STREAM_END) {
 | 
						|
        return -1;
 | 
						|
    }
 | 
						|
 | 
						|
    return stream->total_out;
 | 
						|
}
 | 
						|
 | 
						|
static void *do_data_decompress(void *opaque)
 | 
						|
{
 | 
						|
    DecompressParam *param = opaque;
 | 
						|
    unsigned long pagesize;
 | 
						|
    uint8_t *des;
 | 
						|
    int len, ret;
 | 
						|
 | 
						|
    qemu_mutex_lock(¶m->mutex);
 | 
						|
    while (!param->quit) {
 | 
						|
        if (param->des) {
 | 
						|
            des = param->des;
 | 
						|
            len = param->len;
 | 
						|
            param->des = 0;
 | 
						|
            qemu_mutex_unlock(¶m->mutex);
 | 
						|
 | 
						|
            pagesize = qemu_target_page_size();
 | 
						|
 | 
						|
            ret = qemu_uncompress_data(¶m->stream, des, pagesize,
 | 
						|
                                       param->compbuf, len);
 | 
						|
            if (ret < 0 && migrate_get_current()->decompress_error_check) {
 | 
						|
                error_report("decompress data failed");
 | 
						|
                qemu_file_set_error(decomp_file, ret);
 | 
						|
            }
 | 
						|
 | 
						|
            qemu_mutex_lock(&decomp_done_lock);
 | 
						|
            param->done = true;
 | 
						|
            qemu_cond_signal(&decomp_done_cond);
 | 
						|
            qemu_mutex_unlock(&decomp_done_lock);
 | 
						|
 | 
						|
            qemu_mutex_lock(¶m->mutex);
 | 
						|
        } else {
 | 
						|
            qemu_cond_wait(¶m->cond, ¶m->mutex);
 | 
						|
        }
 | 
						|
    }
 | 
						|
    qemu_mutex_unlock(¶m->mutex);
 | 
						|
 | 
						|
    return NULL;
 | 
						|
}
 | 
						|
 | 
						|
int wait_for_decompress_done(void)
 | 
						|
{
 | 
						|
    int idx, thread_count;
 | 
						|
 | 
						|
    if (!migrate_compress()) {
 | 
						|
        return 0;
 | 
						|
    }
 | 
						|
 | 
						|
    thread_count = migrate_decompress_threads();
 | 
						|
    qemu_mutex_lock(&decomp_done_lock);
 | 
						|
    for (idx = 0; idx < thread_count; idx++) {
 | 
						|
        while (!decomp_param[idx].done) {
 | 
						|
            qemu_cond_wait(&decomp_done_cond, &decomp_done_lock);
 | 
						|
        }
 | 
						|
    }
 | 
						|
    qemu_mutex_unlock(&decomp_done_lock);
 | 
						|
    return qemu_file_get_error(decomp_file);
 | 
						|
}
 | 
						|
 | 
						|
void compress_threads_load_cleanup(void)
 | 
						|
{
 | 
						|
    int i, thread_count;
 | 
						|
 | 
						|
    if (!migrate_compress()) {
 | 
						|
        return;
 | 
						|
    }
 | 
						|
    thread_count = migrate_decompress_threads();
 | 
						|
    for (i = 0; i < thread_count; i++) {
 | 
						|
        /*
 | 
						|
         * we use it as a indicator which shows if the thread is
 | 
						|
         * properly init'd or not
 | 
						|
         */
 | 
						|
        if (!decomp_param[i].compbuf) {
 | 
						|
            break;
 | 
						|
        }
 | 
						|
 | 
						|
        qemu_mutex_lock(&decomp_param[i].mutex);
 | 
						|
        decomp_param[i].quit = true;
 | 
						|
        qemu_cond_signal(&decomp_param[i].cond);
 | 
						|
        qemu_mutex_unlock(&decomp_param[i].mutex);
 | 
						|
    }
 | 
						|
    for (i = 0; i < thread_count; i++) {
 | 
						|
        if (!decomp_param[i].compbuf) {
 | 
						|
            break;
 | 
						|
        }
 | 
						|
 | 
						|
        qemu_thread_join(decompress_threads + i);
 | 
						|
        qemu_mutex_destroy(&decomp_param[i].mutex);
 | 
						|
        qemu_cond_destroy(&decomp_param[i].cond);
 | 
						|
        inflateEnd(&decomp_param[i].stream);
 | 
						|
        g_free(decomp_param[i].compbuf);
 | 
						|
        decomp_param[i].compbuf = NULL;
 | 
						|
    }
 | 
						|
    g_free(decompress_threads);
 | 
						|
    g_free(decomp_param);
 | 
						|
    decompress_threads = NULL;
 | 
						|
    decomp_param = NULL;
 | 
						|
    decomp_file = NULL;
 | 
						|
}
 | 
						|
 | 
						|
int compress_threads_load_setup(QEMUFile *f)
 | 
						|
{
 | 
						|
    int i, thread_count;
 | 
						|
 | 
						|
    if (!migrate_compress()) {
 | 
						|
        return 0;
 | 
						|
    }
 | 
						|
 | 
						|
    thread_count = migrate_decompress_threads();
 | 
						|
    decompress_threads = g_new0(QemuThread, thread_count);
 | 
						|
    decomp_param = g_new0(DecompressParam, thread_count);
 | 
						|
    qemu_mutex_init(&decomp_done_lock);
 | 
						|
    qemu_cond_init(&decomp_done_cond);
 | 
						|
    decomp_file = f;
 | 
						|
    for (i = 0; i < thread_count; i++) {
 | 
						|
        if (inflateInit(&decomp_param[i].stream) != Z_OK) {
 | 
						|
            goto exit;
 | 
						|
        }
 | 
						|
 | 
						|
        size_t compbuf_size = compressBound(qemu_target_page_size());
 | 
						|
        decomp_param[i].compbuf = g_malloc0(compbuf_size);
 | 
						|
        qemu_mutex_init(&decomp_param[i].mutex);
 | 
						|
        qemu_cond_init(&decomp_param[i].cond);
 | 
						|
        decomp_param[i].done = true;
 | 
						|
        decomp_param[i].quit = false;
 | 
						|
        qemu_thread_create(decompress_threads + i, "decompress",
 | 
						|
                           do_data_decompress, decomp_param + i,
 | 
						|
                           QEMU_THREAD_JOINABLE);
 | 
						|
    }
 | 
						|
    return 0;
 | 
						|
exit:
 | 
						|
    compress_threads_load_cleanup();
 | 
						|
    return -1;
 | 
						|
}
 | 
						|
 | 
						|
void decompress_data_with_multi_threads(QEMUFile *f, void *host, int len)
 | 
						|
{
 | 
						|
    int idx, thread_count;
 | 
						|
 | 
						|
    thread_count = migrate_decompress_threads();
 | 
						|
    QEMU_LOCK_GUARD(&decomp_done_lock);
 | 
						|
    while (true) {
 | 
						|
        for (idx = 0; idx < thread_count; idx++) {
 | 
						|
            if (decomp_param[idx].done) {
 | 
						|
                decomp_param[idx].done = false;
 | 
						|
                qemu_mutex_lock(&decomp_param[idx].mutex);
 | 
						|
                qemu_get_buffer(f, decomp_param[idx].compbuf, len);
 | 
						|
                decomp_param[idx].des = host;
 | 
						|
                decomp_param[idx].len = len;
 | 
						|
                qemu_cond_signal(&decomp_param[idx].cond);
 | 
						|
                qemu_mutex_unlock(&decomp_param[idx].mutex);
 | 
						|
                break;
 | 
						|
            }
 | 
						|
        }
 | 
						|
        if (idx < thread_count) {
 | 
						|
            break;
 | 
						|
        } else {
 | 
						|
            qemu_cond_wait(&decomp_done_cond, &decomp_done_lock);
 | 
						|
        }
 | 
						|
    }
 | 
						|
}
 |