Message ID | 20171013181913.7556-1-Andrew.Baumann@microsoft.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
> From: Andrew Baumann > Sent: Friday, 13 October 2017 11:19 > > Aligned 8-byte memory writes by a 64-bit target on a 64-bit host should > always turn into atomic 8-byte writes on the host, however if we missed > in the softmmu, and the TLB line was marked as not dirty, then we > would end up tearing the 8-byte write into two 4-byte writes in > access_with_adjusted_size(). > > Signed-off-by: Andrew Baumann <Andrew.Baumann@microsoft.com> > --- > This manifested as a race in lock-free synchronisation with an aarch64 > Windows guest on an x86-64 host (with multithreaded TCG). > > exec.c | 13 +++++++++++++ > 1 file changed, 13 insertions(+) By the way, I noticed that watch_mem_ops are also 4-byte only. I suspect the same fix may be needed there? Andrew
On 10/13/2017 11:19 AM, Andrew Baumann wrote: > Aligned 8-byte memory writes by a 64-bit target on a 64-bit host should > always turn into atomic 8-byte writes on the host, however if we missed > in the softmmu, and the TLB line was marked as not dirty, then we > would end up tearing the 8-byte write into two 4-byte writes in > access_with_adjusted_size(). > > Signed-off-by: Andrew Baumann <Andrew.Baumann@microsoft.com> > --- > This manifested as a race in lock-free synchronisation with an aarch64 > Windows guest on an x86-64 host (with multithreaded TCG). Reviewed-by: Richard Henderson <richard.henderson@linaro.org> r~
On 10/13/2017 11:27 AM, Andrew Baumann wrote: >> From: Andrew Baumann >> Sent: Friday, 13 October 2017 11:19 >> >> Aligned 8-byte memory writes by a 64-bit target on a 64-bit host should >> always turn into atomic 8-byte writes on the host, however if we missed >> in the softmmu, and the TLB line was marked as not dirty, then we >> would end up tearing the 8-byte write into two 4-byte writes in >> access_with_adjusted_size(). >> >> Signed-off-by: Andrew Baumann <Andrew.Baumann@microsoft.com> >> --- >> This manifested as a race in lock-free synchronisation with an aarch64 >> Windows guest on an x86-64 host (with multithreaded TCG). >> >> exec.c | 13 +++++++++++++ >> 1 file changed, 13 insertions(+) > > By the way, I noticed that watch_mem_ops are also 4-byte only. I suspect the > same fix may be needed there? Yep. r~
diff --git a/exec.c b/exec.c index 6378714a2b..7c591a9b75 100644 --- a/exec.c +++ b/exec.c @@ -2348,6 +2348,9 @@ static void notdirty_mem_write(void *opaque, hwaddr ram_addr, case 4: stl_p(qemu_map_ram_ptr(NULL, ram_addr), val); break; + case 8: + stq_p(qemu_map_ram_ptr(NULL, ram_addr), val); + break; default: abort(); } @@ -2378,6 +2381,16 @@ static const MemoryRegionOps notdirty_mem_ops = { .write = notdirty_mem_write, .valid.accepts = notdirty_mem_accepts, .endianness = DEVICE_NATIVE_ENDIAN, + .valid = { + .min_access_size = 1, + .max_access_size = 8, + .unaligned = false, + }, + .impl = { + .min_access_size = 1, + .max_access_size = 8, + .unaligned = false, + }, }; /* Generate a debug exception if a watchpoint has been hit. */
Aligned 8-byte memory writes by a 64-bit target on a 64-bit host should always turn into atomic 8-byte writes on the host, however if we missed in the softmmu, and the TLB line was marked as not dirty, then we would end up tearing the 8-byte write into two 4-byte writes in access_with_adjusted_size(). Signed-off-by: Andrew Baumann <Andrew.Baumann@microsoft.com> --- This manifested as a race in lock-free synchronisation with an aarch64 Windows guest on an x86-64 host (with multithreaded TCG). exec.c | 13 +++++++++++++ 1 file changed, 13 insertions(+)