@@ -105,10 +105,10 @@ struct qemu_work_item {
static void queue_work_on_cpu(CPUState *cpu, struct qemu_work_item *wi)
{
- qemu_mutex_lock(&cpu->work_mutex);
+ qemu_mutex_lock(&cpu->lock);
QSIMPLEQ_INSERT_TAIL(&cpu->work_list, wi, node);
wi->done = false;
- qemu_mutex_unlock(&cpu->work_mutex);
+ qemu_mutex_unlock(&cpu->lock);
qemu_cpu_kick(cpu);
}
@@ -300,15 +300,15 @@ void process_queued_cpu_work(CPUState *cpu)
{
struct qemu_work_item *wi;
- qemu_mutex_lock(&cpu->work_mutex);
+ qemu_mutex_lock(&cpu->lock);
if (QSIMPLEQ_EMPTY(&cpu->work_list)) {
- qemu_mutex_unlock(&cpu->work_mutex);
+ qemu_mutex_unlock(&cpu->lock);
return;
}
while (!QSIMPLEQ_EMPTY(&cpu->work_list)) {
wi = QSIMPLEQ_FIRST(&cpu->work_list);
QSIMPLEQ_REMOVE_HEAD(&cpu->work_list, node);
- qemu_mutex_unlock(&cpu->work_mutex);
+ qemu_mutex_unlock(&cpu->lock);
if (wi->exclusive) {
/* Running work items outside the BQL avoids the following deadlock:
* 1) start_exclusive() is called with the BQL taken while another
@@ -324,13 +324,13 @@ void process_queued_cpu_work(CPUState *cpu)
} else {
wi->func(cpu, wi->data);
}
- qemu_mutex_lock(&cpu->work_mutex);
+ qemu_mutex_lock(&cpu->lock);
if (wi->free) {
g_free(wi);
} else {
atomic_mb_set(&wi->done, true);
}
}
- qemu_mutex_unlock(&cpu->work_mutex);
+ qemu_mutex_unlock(&cpu->lock);
qemu_cond_broadcast(&qemu_work_cond);
}
@@ -101,9 +101,9 @@ static inline bool cpu_work_list_empty(CPUState *cpu)
{
bool ret;
- qemu_mutex_lock(&cpu->work_mutex);
+ qemu_mutex_lock(&cpu->lock);
ret = QSIMPLEQ_EMPTY(&cpu->work_list);
- qemu_mutex_unlock(&cpu->work_mutex);
+ qemu_mutex_unlock(&cpu->lock);
return ret;
}
@@ -367,7 +367,7 @@ static void cpu_common_initfn(Object *obj)
cpu->nr_cores = 1;
cpu->nr_threads = 1;
- qemu_mutex_init(&cpu->work_mutex);
+ qemu_mutex_init(&cpu->lock);
QSIMPLEQ_INIT(&cpu->work_list);
QTAILQ_INIT(&cpu->breakpoints);
QTAILQ_INIT(&cpu->watchpoints);
@@ -379,7 +379,7 @@ static void cpu_common_finalize(Object *obj)
{
CPUState *cpu = CPU(obj);
- qemu_mutex_destroy(&cpu->work_mutex);
+ qemu_mutex_destroy(&cpu->lock);
}
static int64_t cpu_common_get_arch_id(CPUState *cpu)
@@ -331,7 +331,8 @@ struct qemu_work_item;
* @opaque: User data.
* @mem_io_pc: Host Program Counter at which the memory was accessed.
* @kvm_fd: vCPU file descriptor for KVM.
- * @work_mutex: Lock to prevent multiple access to @work_list.
+ * @lock: Lock to prevent multiple access to per-CPU fields. Must be acquired
+ * after the BQL.
* @work_list: List of pending asynchronous work.
* @trace_dstate_delayed: Delayed changes to trace_dstate (includes all changes
* to @trace_dstate).
@@ -375,7 +376,8 @@ struct CPUState {
uint64_t random_seed;
sigjmp_buf jmp_env;
- QemuMutex work_mutex;
+ QemuMutex lock;
+ /* fields below protected by @lock */
QSIMPLEQ_HEAD(, qemu_work_item) work_list;
CPUAddressSpace *cpu_ases;