Message ID | 20241017143211.17771-5-farosas@suse.de (mailing list archive) |
---|---|
State | New |
Headers | show |
Series | tests/qtest: Move the bulk of migration tests into a separate target | expand |
On Thu, Oct 17, 2024 at 11:32:11AM -0300, Fabiano Rosas wrote: > Recent changes to how we invoke the migration tests have > (intentionally) caused them to not be part of the check-qtest target > anymore. Add the check-migration-quick target so we don't lose > migration code testing in this job. But 'check-migration-quick' is only the subset of migration tests, 'check-migration' is all of the migration tests. So surely this is a massive regressions in covage in CI pipelines. Experience shows us that relying on humans to run tests periodically doesn't work well, and they'll slowly bit rot. Migration maintainers don't have a way to run this as gating test for every pull request that merges, and pull requests coming from non-migration maintainers can still break migration code. Any tests in tree need to be exercised by CI as the minimum bar to prevent bit rot from merges. > > Signed-off-by: Fabiano Rosas <farosas@suse.de> > --- > .gitlab-ci.d/buildtest.yml | 2 +- > 1 file changed, 1 insertion(+), 1 deletion(-) > > diff --git a/.gitlab-ci.d/buildtest.yml b/.gitlab-ci.d/buildtest.yml > index 34d3f4e9ab..37247dc5fa 100644 > --- a/.gitlab-ci.d/buildtest.yml > +++ b/.gitlab-ci.d/buildtest.yml > @@ -442,7 +442,7 @@ clang-system: > CONFIGURE_ARGS: --cc=clang --cxx=clang++ --enable-ubsan > --extra-cflags=-fno-sanitize-recover=undefined > TARGETS: alpha-softmmu arm-softmmu m68k-softmmu mips64-softmmu s390x-softmmu > - MAKE_CHECK_ARGS: check-qtest check-tcg > + MAKE_CHECK_ARGS: check-qtest check-tcg check-migration-quick > > clang-user: > extends: .native_build_job_template > -- > 2.35.3 > > With regards, Daniel
Daniel P. Berrangé <berrange@redhat.com> writes: > On Thu, Oct 17, 2024 at 11:32:11AM -0300, Fabiano Rosas wrote: >> Recent changes to how we invoke the migration tests have >> (intentionally) caused them to not be part of the check-qtest target >> anymore. Add the check-migration-quick target so we don't lose >> migration code testing in this job. > > But 'check-migration-quick' is only the subset of migration tests, > 'check-migration' is all of the migration tests. So surely this is > a massive regressions in covage in CI pipelines. I'm not sure it is. There are tests there already for all the major parts of the code: precopy, postcopy, multifd, socket. Besides, we can tweak migration-quick to cover spots where we think we're losing coverage. Since our CI offers nothing in terms of reproducibility or debuggability, I don't think it's productive to have an increasing amount of tests running in CI if that means we'll be dealing with timeouts and intermittent crashes constantly. > > Experience shows us that relying on humans to run tests periodically > doesn't work well, and they'll slowly bit rot. Migration maintainers > don't have a way to run this as gating test for every pull request > that merges, and pull requests coming from non-migration maintainers > can still break migration code. Right, but migration code would still be tested with migration-quick which is executed at every make check. Do we really need the full set in every pull request? We must draw a line somewhere, otherwise make check will just balloon in duration. > > Any tests in tree need to be exercised by CI as the minimum bar > to prevent bit rot from merges. > No disagreement here. But then I'm going to need advice on what to do when other maintainers ask us to stop writing migration tests because they take too long. I cannot send contributors away nor merge code without tests. Do we need nightly CI runs? Unit tests? Bear in mind there's a resource allocation issue there. Addressing problems with timeouts/races in our CI is not something any random person can do. >> >> Signed-off-by: Fabiano Rosas <farosas@suse.de> >> --- >> .gitlab-ci.d/buildtest.yml | 2 +- >> 1 file changed, 1 insertion(+), 1 deletion(-) >> >> diff --git a/.gitlab-ci.d/buildtest.yml b/.gitlab-ci.d/buildtest.yml >> index 34d3f4e9ab..37247dc5fa 100644 >> --- a/.gitlab-ci.d/buildtest.yml >> +++ b/.gitlab-ci.d/buildtest.yml >> @@ -442,7 +442,7 @@ clang-system: >> CONFIGURE_ARGS: --cc=clang --cxx=clang++ --enable-ubsan >> --extra-cflags=-fno-sanitize-recover=undefined >> TARGETS: alpha-softmmu arm-softmmu m68k-softmmu mips64-softmmu s390x-softmmu >> - MAKE_CHECK_ARGS: check-qtest check-tcg >> + MAKE_CHECK_ARGS: check-qtest check-tcg check-migration-quick >> >> clang-user: >> extends: .native_build_job_template >> -- >> 2.35.3 >> >> > > With regards, > Daniel
On Thu, Oct 17, 2024 at 01:29:35PM -0300, Fabiano Rosas wrote: > Daniel P. Berrangé <berrange@redhat.com> writes: > > > On Thu, Oct 17, 2024 at 11:32:11AM -0300, Fabiano Rosas wrote: > >> Recent changes to how we invoke the migration tests have > >> (intentionally) caused them to not be part of the check-qtest target > >> anymore. Add the check-migration-quick target so we don't lose > >> migration code testing in this job. > > > > But 'check-migration-quick' is only the subset of migration tests, > > 'check-migration' is all of the migration tests. So surely this is > > a massive regressions in covage in CI pipelines. > > I'm not sure it is. There are tests there already for all the major > parts of the code: precopy, postcopy, multifd, socket. Besides, we can > tweak migration-quick to cover spots where we think we're losing > coverage. Each of the tests in migration-test were added for a good reason, generally to address testing gaps where we had functional regressions in the past. I don't think its a good idea to stop running such tests in CI as gating on new contributions. Any time we've had optional tests in QEMU, we've seen repeated regressions in the area in question. > Since our CI offers nothing in terms of reproducibility or > debuggability, I don't think it's productive to have an increasing > amount of tests running in CI if that means we'll be dealing with > timeouts and intermittent crashes constantly. Test reliability is a different thing. If a particular test is flaky, it needs to either be fixed or disabled. Splitting into a fast & slow grouping doesn't address reliability, just hides the problem from view. > > Experience shows us that relying on humans to run tests periodically > > doesn't work well, and they'll slowly bit rot. Migration maintainers > > don't have a way to run this as gating test for every pull request > > that merges, and pull requests coming from non-migration maintainers > > can still break migration code. > > Right, but migration code would still be tested with migration-quick > which is executed at every make check. Do we really need the full set in > every pull request? We must draw a line somewhere, otherwise make check > will just balloon in duration. Again, the tests all exist because migration code is incredibly complicated, with a lot of permutations, with a history of being very bug / regression prone. With that in mind, it is unavoidable that we're going to have a significant testing overhead for migration code. Looking at its execution time right now, I'd say migration test is pretty good, considering the permutations we have to target. It gets a bad reputation because historically it has been as much as x20 slower than it is today, and has also struggled with reliability. The latter is a reflection of the complexity of migration and and IMHO actually justifies greater testing, as long as we put in time to address bugs. Also we've got one single test program, covering an entire subsystem in one go, rather than lots of short individual test programs, so migration unfairly gets blamed for being slow, when it simply covers alot of functionality in one program. > > Any tests in tree need to be exercised by CI as the minimum bar > > to prevent bit rot from merges. > > > > No disagreement here. But then I'm going to need advice on what to do > when other maintainers ask us to stop writing migration tests because > they take too long. I cannot send contributors away nor merge code > without tests. In general, I think it is unreasonable for other maintainers to tell us to stop adding test coverage for migration, and would push back against such a request. We should, however, continue to optimize how we add further test coverage, where practical, overload testing of multiple features onto a single test case helps. We've already massively optimized the migration-test compared to its historical behaviour. A potentially bigger win could be seen if we change how we exercise the migration functionality. Since we had the migration qtest that runs a full migration operation, we've tended to expand testing by adding new qtest functions. ie we've added a functional test for everything we want covered. This is nice & simple, but also expensive. We've ignored unit testing, which I think is a mistake. If i look at the test list: # /x86_64/migration/bad_dest # /x86_64/migration/analyze-script # /x86_64/migration/validate_uuid # /x86_64/migration/validate_uuid_error # /x86_64/migration/validate_uuid_src_not_set # /x86_64/migration/validate_uuid_dst_not_set # /x86_64/migration/dirty_ring # /x86_64/migration/precopy/file # /x86_64/migration/precopy/unix/plain # /x86_64/migration/precopy/unix/suspend/live # /x86_64/migration/precopy/unix/suspend/notlive # /x86_64/migration/precopy/unix/tls/psk # /x86_64/migration/precopy/unix/tls/x509/default-host # /x86_64/migration/precopy/unix/tls/x509/override-host # /x86_64/migration/precopy/file/offset # /x86_64/migration/precopy/file/mapped-ram # /x86_64/migration/precopy/file/offset/fdset # /x86_64/migration/precopy/file/offset/bad # /x86_64/migration/precopy/file/mapped-ram/live # /x86_64/migration/precopy/tcp/plain # /x86_64/migration/precopy/tcp/plain/switchover-ack # /x86_64/migration/precopy/tcp/tls/psk/match # /x86_64/migration/precopy/tcp/tls/psk/mismatch # /x86_64/migration/precopy/tcp/tls/x509/default-host # /x86_64/migration/precopy/tcp/tls/x509/override-host # /x86_64/migration/precopy/tcp/tls/x509/mismatch-host # /x86_64/migration/precopy/tcp/tls/x509/friendly-client # /x86_64/migration/precopy/tcp/tls/x509/hostile-client # /x86_64/migration/precopy/tcp/tls/x509/allow-anon-client # /x86_64/migration/precopy/tcp/tls/x509/reject-anon-client # /x86_64/migration/precopy/fd/tcp # /x86_64/migration/precopy/fd/file # /x86_64/migration/multifd/file/mapped-ram # /x86_64/migration/multifd/file/mapped-ram/live # /x86_64/migration/multifd/file/mapped-ram/dio # /x86_64/migration/multifd/file/mapped-ram/fdset # /x86_64/migration/multifd/file/mapped-ram/fdset/dio # /x86_64/migration/multifd/tcp/uri/plain/none # /x86_64/migration/multifd/tcp/channels/plain/none # /x86_64/migration/multifd/tcp/plain/cancel # /x86_64/migration/multifd/tcp/plain/zlib # /x86_64/migration/multifd/tcp/plain/zstd # /x86_64/migration/multifd/tcp/plain/zero-page/legacy # /x86_64/migration/multifd/tcp/plain/zero-page/none # /x86_64/migration/multifd/tcp/tls/psk/match # /x86_64/migration/multifd/tcp/tls/psk/mismatch # /x86_64/migration/multifd/tcp/tls/x509/default-host # /x86_64/migration/multifd/tcp/tls/x509/override-host # /x86_64/migration/multifd/tcp/tls/x509/mismatch-host # /x86_64/migration/multifd/tcp/tls/x509/allow-anon-client # /x86_64/migration/multifd/tcp/tls/x509/reject-anon-client # /x86_64/migration/validate_uri/channels/both_set # /x86_64/migration/validate_uri/channels/none_set Individually none of those is very slow on its own - 10 are in the 2-3 second range, 35 are 1-2 secs, and 6 are less than 1 second. A very large portion of those are validating different ways to establish migration. Hardly any of them actually need to run a migration to completion. Even without running to completion though, we have the overheads of spawning 2 QEMUs. This feels like something that should be amenable to unit testing. Might need a little re-factoring of migration code to make it easier to run a subset of its logic in isolation, but that'd probably be a win anyway, as such work usually makes code cleaner. > Do we need nightly CI runs? Unit tests? Bear in mind there's a resource > allocation issue there. Addressing problems with timeouts/races in our > CI is not something any random person can do. In terms of running time, I think migration-test is acceptable as it is to run in 'make check' by default and doesn't justify dropping test coverage. We should still look to optimize & move to unit testing more code, and any reliability issues are something that needs to be addressed too. With regards, Daniel
On Fri, 18 Oct 2024 at 10:01, Daniel P. Berrangé <berrange@redhat.com> wrote: > > On Thu, Oct 17, 2024 at 01:29:35PM -0300, Fabiano Rosas wrote: > > Daniel P. Berrangé <berrange@redhat.com> writes: > > > > > On Thu, Oct 17, 2024 at 11:32:11AM -0300, Fabiano Rosas wrote: > > >> Recent changes to how we invoke the migration tests have > > >> (intentionally) caused them to not be part of the check-qtest target > > >> anymore. Add the check-migration-quick target so we don't lose > > >> migration code testing in this job. > > > > > > But 'check-migration-quick' is only the subset of migration tests, > > > 'check-migration' is all of the migration tests. So surely this is > > > a massive regressions in covage in CI pipelines. > > > > I'm not sure it is. There are tests there already for all the major > > parts of the code: precopy, postcopy, multifd, socket. Besides, we can > > tweak migration-quick to cover spots where we think we're losing > > coverage. > > Each of the tests in migration-test were added for a good reason, > generally to address testing gaps where we had functional regressions > in the past. I don't think its a good idea to stop running such tests > in CI as gating on new contributions. Any time we've had optional > tests in QEMU, we've seen repeated regressions in the area in question. > > > Since our CI offers nothing in terms of reproducibility or > > debuggability, I don't think it's productive to have an increasing > > amount of tests running in CI if that means we'll be dealing with > > timeouts and intermittent crashes constantly. > > Test reliability is a different thing. If a particular test is > flaky, it needs to either be fixed or disabled. Splitting into > a fast & slow grouping doesn't address reliability, just hides > the problem from view. A lot of the current reliability issue is timeouts -- sometimes our CI runners just run really slow (I have seen an example where between a normal and a slow run on the same commit both the compile and test times were 10x different...) So any test that is not a fast-to-complete is much much more likely to hit its timeout if the runner is running slowly. When I am doing CI testing for merges "migration test timed out again" is really really common. > > No disagreement here. But then I'm going to need advice on what to do > > when other maintainers ask us to stop writing migration tests because > > they take too long. I cannot send contributors away nor merge code > > without tests. > > In general, I think it is unreasonable for other maintainers to > tell us to stop adding test coverage for migration, and would > push back against such a request. We do not have infinite CI resources, unfortunately. Migration is competing with everything else for time on CI. You have to find a balance between "what do we run every time" and "what do we only run when specifically testing a migration pullreq". Similarly, there's a lot of iotests but we don't run all of them for every block backend for every CI job via "make check". Long test times for tests run under "make check" are also bad for individual developers -- if I'm running "make check" to test a target/arm change I've made I don't really want that to then spend 15 minutes testing the migration code that I haven't touched and that is vanishingly unlikely to be affected by my patches. thanks -- PMM
On Fri, Oct 18, 2024 at 10:46:55AM +0100, Peter Maydell wrote: > On Fri, 18 Oct 2024 at 10:01, Daniel P. Berrangé <berrange@redhat.com> wrote: > > > > On Thu, Oct 17, 2024 at 01:29:35PM -0300, Fabiano Rosas wrote: > > > Daniel P. Berrangé <berrange@redhat.com> writes: > > > > > > > On Thu, Oct 17, 2024 at 11:32:11AM -0300, Fabiano Rosas wrote: > > > >> Recent changes to how we invoke the migration tests have > > > >> (intentionally) caused them to not be part of the check-qtest target > > > >> anymore. Add the check-migration-quick target so we don't lose > > > >> migration code testing in this job. > > > > > > > > But 'check-migration-quick' is only the subset of migration tests, > > > > 'check-migration' is all of the migration tests. So surely this is > > > > a massive regressions in covage in CI pipelines. > > > > > > I'm not sure it is. There are tests there already for all the major > > > parts of the code: precopy, postcopy, multifd, socket. Besides, we can > > > tweak migration-quick to cover spots where we think we're losing > > > coverage. > > > > Each of the tests in migration-test were added for a good reason, > > generally to address testing gaps where we had functional regressions > > in the past. I don't think its a good idea to stop running such tests > > in CI as gating on new contributions. Any time we've had optional > > tests in QEMU, we've seen repeated regressions in the area in question. > > > > > Since our CI offers nothing in terms of reproducibility or > > > debuggability, I don't think it's productive to have an increasing > > > amount of tests running in CI if that means we'll be dealing with > > > timeouts and intermittent crashes constantly. > > > > Test reliability is a different thing. If a particular test is > > flaky, it needs to either be fixed or disabled. Splitting into > > a fast & slow grouping doesn't address reliability, just hides > > the problem from view. > > A lot of the current reliability issue is timeouts -- sometimes > our CI runners just run really slow (I have seen an example where > between a normal and a slow run on the same commit both the > compile and test times were 10x different...) So any test > that is not a fast-to-complete is much much more likely to > hit its timeout if the runner is running slowly. When I am > doing CI testing for merges "migration test timed out again" > is really really common. If its frequently timing out, then we've got the timeouts wrong, or we have some genuine bugs in there to be fixed. > > > No disagreement here. But then I'm going to need advice on what to do > > > when other maintainers ask us to stop writing migration tests because > > > they take too long. I cannot send contributors away nor merge code > > > without tests. > > > > In general, I think it is unreasonable for other maintainers to > > tell us to stop adding test coverage for migration, and would > > push back against such a request. > > We do not have infinite CI resources, unfortunately. Migration > is competing with everything else for time on CI. You have to > find a balance between "what do we run every time" and "what > do we only run when specifically testing a migration pullreq". > Similarly, there's a lot of iotests but we don't run all of them > for every block backend for every CI job via "make check". The combos we don't run for iotests are a good source of regressions too :-( > Long test times for tests run under "make check" are also bad > for individual developers -- if I'm running "make check" to > test a target/arm change I've made I don't really want that > to then spend 15 minutes testing the migration code that > I haven't touched and that is vanishingly unlikely to be > affected by my patches. Migration-test *used* to take 15 minutes to run, but that was a very long time ago. A run of it today is around 1m20. That said, if you are building multiple system emulators, we run the same test multiple times, and with the number of targets we have, that will be painful. That could be a good reason to split the migration-test into two distinct programs. One program that runs for every target, and one that is only run once, for some arbitrary "primary" target ? Or could we make use of glib's g_test_thorough for this - a primary target runs with "SPEED=through" and all other targets with normal settings. That would give us a way to optimize any of the qtests to reduce redundant testing where appropriate. If we move alot of testing out into a migration unit test, this also solves the redundancy problem. With regards, Daniel
On Fri, 18 Oct 2024 at 11:00, Daniel P. Berrangé <berrange@redhat.com> wrote: > > On Fri, Oct 18, 2024 at 10:46:55AM +0100, Peter Maydell wrote: > > We do not have infinite CI resources, unfortunately. Migration > > is competing with everything else for time on CI. You have to > > find a balance between "what do we run every time" and "what > > do we only run when specifically testing a migration pullreq". > > Similarly, there's a lot of iotests but we don't run all of them > > for every block backend for every CI job via "make check". > > The combos we don't run for iotests are a good source of > regressions too :-( > > > Long test times for tests run under "make check" are also bad > > for individual developers -- if I'm running "make check" to > > test a target/arm change I've made I don't really want that > > to then spend 15 minutes testing the migration code that > > I haven't touched and that is vanishingly unlikely to be > > affected by my patches. > > Migration-test *used* to take 15 minutes to run, but that was a > very long time ago. A run of it today is around 1m20. > > That said, if you are building multiple system emulators, we > run the same test multiple times, and with the number of > targets we have, that will be painful. Yeah. Here's a recent s390 job, and not one that was running slow: https://gitlab.com/qemu-project/qemu/-/jobs/8112195449 The migration tests it ran were: 95/954 qemu:qtest+qtest-ppc64 / qtest-ppc64/migration-test OK 196.95s 50 subtests passed 96/954 qemu:qtest+qtest-aarch64 / qtest-aarch64/migration-test OK 202.47s 50 subtests passed 99/954 qemu:qtest+qtest-i386 / qtest-i386/migration-test OK 203.54s 52 subtests passed 107/954 qemu:qtest+qtest-s390x / qtest-s390x/migration-test OK 193.65s 50 subtests passed 156/954 qemu:qtest+qtest-x86_64 / qtest-x86_64/migration-test OK 164.44s 52 subtests passed So that's 192s (over 3 minutes) average, and over 16 minutes total spent on migration testing on this one CI job. If the s390 VM has a noisy-neighbour problem then the migration tests can hit their 8 minute timeout, implying 40 minutes spent on migration testing alone... thanks -- PMM
Daniel P. Berrangé <berrange@redhat.com> writes: > On Fri, Oct 18, 2024 at 10:46:55AM +0100, Peter Maydell wrote: >> On Fri, 18 Oct 2024 at 10:01, Daniel P. Berrangé <berrange@redhat.com> wrote: >> > >> > On Thu, Oct 17, 2024 at 01:29:35PM -0300, Fabiano Rosas wrote: >> > > Daniel P. Berrangé <berrange@redhat.com> writes: >> > > >> > > > On Thu, Oct 17, 2024 at 11:32:11AM -0300, Fabiano Rosas wrote: <snip> > > Migration-test *used* to take 15 minutes to run, but that was a > very long time ago. A run of it today is around 1m20. > > That said, if you are building multiple system emulators, we > run the same test multiple times, and with the number of > targets we have, that will be painful. I think this is the main problem. We run the whole set of tests for every system emulator we build and I doubt we are actually increasing any coverage. One suggestion I made previously was to change the test selection logic so we only run all the tests for the KVM target and one TCG target. For the other targets we should run the basic tests only to ensure there is now architecture specific breakage for migration generally. > That could be a good reason to split the migration-test into > two distinct programs. One program that runs for every target, > and one that is only run once, for some arbitrary "primary" > target ? Or could we make use of glib's g_test_thorough > for this - a primary target runs with "SPEED=through" and > all other targets with normal settings. That would give us > a way to optimize any of the qtests to reduce redundant > testing where appropriate. Does splitting the tests make it easier? Would meson just pick which arches for the "migrtion-full" experience? My idea was just to pass the list of configured builds as an environment variable and without it default to everything. That way running the test by hand would still have full coverage. > > > If we move alot of testing out into a migration unit test, > this also solves the redundancy problem. > > > With regards, > Daniel
Daniel P. Berrangé <berrange@redhat.com> writes: > On Fri, Oct 18, 2024 at 10:46:55AM +0100, Peter Maydell wrote: >> On Fri, 18 Oct 2024 at 10:01, Daniel P. Berrangé <berrange@redhat.com> wrote: >> > >> > On Thu, Oct 17, 2024 at 01:29:35PM -0300, Fabiano Rosas wrote: >> > > Daniel P. Berrangé <berrange@redhat.com> writes: >> > > >> > > > On Thu, Oct 17, 2024 at 11:32:11AM -0300, Fabiano Rosas wrote: >> > > >> Recent changes to how we invoke the migration tests have >> > > >> (intentionally) caused them to not be part of the check-qtest target >> > > >> anymore. Add the check-migration-quick target so we don't lose >> > > >> migration code testing in this job. >> > > > >> > > > But 'check-migration-quick' is only the subset of migration tests, >> > > > 'check-migration' is all of the migration tests. So surely this is >> > > > a massive regressions in covage in CI pipelines. >> > > >> > > I'm not sure it is. There are tests there already for all the major >> > > parts of the code: precopy, postcopy, multifd, socket. Besides, we can >> > > tweak migration-quick to cover spots where we think we're losing >> > > coverage. >> > >> > Each of the tests in migration-test were added for a good reason, >> > generally to address testing gaps where we had functional regressions >> > in the past. I don't think its a good idea to stop running such tests >> > in CI as gating on new contributions. Any time we've had optional >> > tests in QEMU, we've seen repeated regressions in the area in question. >> > >> > > Since our CI offers nothing in terms of reproducibility or >> > > debuggability, I don't think it's productive to have an increasing >> > > amount of tests running in CI if that means we'll be dealing with >> > > timeouts and intermittent crashes constantly. >> > >> > Test reliability is a different thing. If a particular test is >> > flaky, it needs to either be fixed or disabled. Splitting into >> > a fast & slow grouping doesn't address reliability, just hides >> > the problem from view. >> >> A lot of the current reliability issue is timeouts -- sometimes >> our CI runners just run really slow (I have seen an example where >> between a normal and a slow run on the same commit both the >> compile and test times were 10x different...) So any test >> that is not a fast-to-complete is much much more likely to >> hit its timeout if the runner is running slowly. When I am >> doing CI testing for merges "migration test timed out again" >> is really really common. > > If its frequently timing out, then we've got the timeouts > wrong, or we have some genuine bugs in there to be fixed. > >> > > No disagreement here. But then I'm going to need advice on what to do >> > > when other maintainers ask us to stop writing migration tests because >> > > they take too long. I cannot send contributors away nor merge code >> > > without tests. >> > >> > In general, I think it is unreasonable for other maintainers to >> > tell us to stop adding test coverage for migration, and would >> > push back against such a request. >> >> We do not have infinite CI resources, unfortunately. Migration >> is competing with everything else for time on CI. You have to >> find a balance between "what do we run every time" and "what >> do we only run when specifically testing a migration pullreq". >> Similarly, there's a lot of iotests but we don't run all of them >> for every block backend for every CI job via "make check". > > The combos we don't run for iotests are a good source of > regressions too :-( > >> Long test times for tests run under "make check" are also bad >> for individual developers -- if I'm running "make check" to >> test a target/arm change I've made I don't really want that >> to then spend 15 minutes testing the migration code that >> I haven't touched and that is vanishingly unlikely to be >> affected by my patches. > > Migration-test *used* to take 15 minutes to run, but that was a > very long time ago. A run of it today is around 1m20. > > That said, if you are building multiple system emulators, we > run the same test multiple times, and with the number of > targets we have, that will be painful. > > That could be a good reason to split the migration-test into > two distinct programs. One program that runs for every target, > and one that is only run once, for some arbitrary "primary" > target ? What do you mean by distinct programs? It's not the migration-test that decides on which targets it runs, it's meson.build. We register a test() for each target, same as with any other qtest. Maybe I misunderstood you... > Or could we make use of glib's g_test_thorough > for this - a primary target runs with "SPEED=through" and > all other targets with normal settings. That would give us > a way to optimize any of the qtests to reduce redundant > testing where appropriate. This still requires a new make target I think. Otherwise we'd run *all* thorough tests for a QEMU target and not only migration-test in thorough mode. > > > If we move alot of testing out into a migration unit test, > this also solves the redundancy problem. > > > With regards, > Daniel
Daniel P. Berrangé <berrange@redhat.com> writes: > On Thu, Oct 17, 2024 at 01:29:35PM -0300, Fabiano Rosas wrote: >> Daniel P. Berrangé <berrange@redhat.com> writes: >> >> > On Thu, Oct 17, 2024 at 11:32:11AM -0300, Fabiano Rosas wrote: >> >> Recent changes to how we invoke the migration tests have >> >> (intentionally) caused them to not be part of the check-qtest target >> >> anymore. Add the check-migration-quick target so we don't lose >> >> migration code testing in this job. >> > >> > But 'check-migration-quick' is only the subset of migration tests, >> > 'check-migration' is all of the migration tests. So surely this is >> > a massive regressions in covage in CI pipelines. >> >> I'm not sure it is. There are tests there already for all the major >> parts of the code: precopy, postcopy, multifd, socket. Besides, we can >> tweak migration-quick to cover spots where we think we're losing >> coverage. > > Each of the tests in migration-test were added for a good reason, > generally to address testing gaps where we had functional regressions > in the past. I don't think its a good idea to stop running such tests > in CI as gating on new contributions. Any time we've had optional > tests in QEMU, we've seen repeated regressions in the area in question. > >> Since our CI offers nothing in terms of reproducibility or >> debuggability, I don't think it's productive to have an increasing >> amount of tests running in CI if that means we'll be dealing with >> timeouts and intermittent crashes constantly. > > Test reliability is a different thing. If a particular test is > flaky, it needs to either be fixed or disabled. The problem is that in this community the idea of "fix" is: wait until someone with the appropriate skill level and interest stumbles upon the problem on their own and fix it in anger. For it to be a proper strategy, we'd need to create an issue in gitlab referencing the bug, have a proper reproducer and encourage contributors to work on the issue. Iff the above was in place, then we could disable the test. Otherwise the test just sits there disabled. In the past there were even tests committed that *never* ran. The situation with multifd/cancel alone is absurd: - It was disabled in March 2023 and stood there *not testing anything* while a major refactoring of the test code was happening. - The test was fixed in June 2023, but not reenabled in fear of getting flak from the community for breaking CI again (or at least that's the feeling I got from talking to Juan). - mapped-ram (which relies entirely on multifd) started being worked on and I had to enable the test in my own branch to be able to test the code properly. While disabled, it caught several issues in mapped-ram. - In October 2023 the test is re-enabled an immediately exposes issues in the code. This is how I started working on the migration code. Maybe you can appreciate why I don't feel confident about this fix or disable strategy. It has eaten many hours of my work. > Splitting into > a fast & slow grouping doesn't address reliability, just hides > the problem from view. Right, and is that not the same as FLAKY? What good is keeping a test in view if the only people that can fix it are the ones that would be seeing the breakage constantly in their own branches anyway? > >> > Experience shows us that relying on humans to run tests periodically >> > doesn't work well, and they'll slowly bit rot. Migration maintainers >> > don't have a way to run this as gating test for every pull request >> > that merges, and pull requests coming from non-migration maintainers >> > can still break migration code. >> >> Right, but migration code would still be tested with migration-quick >> which is executed at every make check. Do we really need the full set in >> every pull request? We must draw a line somewhere, otherwise make check >> will just balloon in duration. > > Again, the tests all exist because migration code is incredibly > complicated, with a lot of permutations, with a history of being > very bug / regression prone. With that in mind, it is unavoidable > that we're going to have a significant testing overhead for > migration code. Yep, we should be testing way more actually. > > Looking at its execution time right now, I'd say migration test > is pretty good, considering the permutations we have to target. > > It gets a bad reputation because historically it has been as > much as x20 slower than it is today, and has also struggled > with reliability. The latter is a reflection of the complexity > of migration and and IMHO actually justifies greater testing, > as long as we put in time to address bugs. > > Also we've got one single test program, covering an entire > subsystem in one go, rather than lots of short individual > test programs, so migration unfairly gets blamed for being > slow, when it simply covers alot of functionality in one > program. And still you think it's not worth it having a separate target for testing migration. FWIW, I also proposed splittling it into multiple meson tests, which you also rejected. It would be so much easier to move all of this into a separate target and let those who want nothing do to with to just ignore it. > >> > Any tests in tree need to be exercised by CI as the minimum bar >> > to prevent bit rot from merges. >> > >> >> No disagreement here. But then I'm going to need advice on what to do >> when other maintainers ask us to stop writing migration tests because >> they take too long. I cannot send contributors away nor merge code >> without tests. > > In general, I think it is unreasonable for other maintainers to > tell us to stop adding test coverage for migration, and would > push back against such a request. This might be a larger issue in QEMU. I also heard the same back in 2021 when doing ppc work: "don't add too many tests because the CI buckles and people get mad". The same with adding too much logging really. We're hostages to the gitlab CI unfortunately. > > We should, however, continue to optimize how we add further test > coverage, where practical, overload testing of multiple features > onto a single test case helps. > > We've already massively optimized the migration-test compared to > its historical behaviour. > > A potentially bigger win could be seen if we change how we exercise > the migration functionality. Since we had the migration qtest that > runs a full migration operation, we've tended to expand testing by > adding new qtest functions. ie we've added a functional test for > everything we want covered. This is nice & simple, but also expensive. > We've ignored unit testing, which I think is a mistake. > > If i look at the test list: > > # /x86_64/migration/bad_dest > # /x86_64/migration/analyze-script > # /x86_64/migration/validate_uuid > # /x86_64/migration/validate_uuid_error > # /x86_64/migration/validate_uuid_src_not_set > # /x86_64/migration/validate_uuid_dst_not_set > # /x86_64/migration/dirty_ring > # /x86_64/migration/precopy/file > # /x86_64/migration/precopy/unix/plain > # /x86_64/migration/precopy/unix/suspend/live > # /x86_64/migration/precopy/unix/suspend/notlive > # /x86_64/migration/precopy/unix/tls/psk > # /x86_64/migration/precopy/unix/tls/x509/default-host > # /x86_64/migration/precopy/unix/tls/x509/override-host > # /x86_64/migration/precopy/file/offset > # /x86_64/migration/precopy/file/mapped-ram > # /x86_64/migration/precopy/file/offset/fdset > # /x86_64/migration/precopy/file/offset/bad > # /x86_64/migration/precopy/file/mapped-ram/live > # /x86_64/migration/precopy/tcp/plain > # /x86_64/migration/precopy/tcp/plain/switchover-ack > # /x86_64/migration/precopy/tcp/tls/psk/match > # /x86_64/migration/precopy/tcp/tls/psk/mismatch > # /x86_64/migration/precopy/tcp/tls/x509/default-host > # /x86_64/migration/precopy/tcp/tls/x509/override-host > # /x86_64/migration/precopy/tcp/tls/x509/mismatch-host > # /x86_64/migration/precopy/tcp/tls/x509/friendly-client > # /x86_64/migration/precopy/tcp/tls/x509/hostile-client > # /x86_64/migration/precopy/tcp/tls/x509/allow-anon-client > # /x86_64/migration/precopy/tcp/tls/x509/reject-anon-client > # /x86_64/migration/precopy/fd/tcp > # /x86_64/migration/precopy/fd/file > # /x86_64/migration/multifd/file/mapped-ram > # /x86_64/migration/multifd/file/mapped-ram/live > # /x86_64/migration/multifd/file/mapped-ram/dio > # /x86_64/migration/multifd/file/mapped-ram/fdset > # /x86_64/migration/multifd/file/mapped-ram/fdset/dio > # /x86_64/migration/multifd/tcp/uri/plain/none > # /x86_64/migration/multifd/tcp/channels/plain/none > # /x86_64/migration/multifd/tcp/plain/cancel > # /x86_64/migration/multifd/tcp/plain/zlib > # /x86_64/migration/multifd/tcp/plain/zstd > # /x86_64/migration/multifd/tcp/plain/zero-page/legacy > # /x86_64/migration/multifd/tcp/plain/zero-page/none > # /x86_64/migration/multifd/tcp/tls/psk/match > # /x86_64/migration/multifd/tcp/tls/psk/mismatch > # /x86_64/migration/multifd/tcp/tls/x509/default-host > # /x86_64/migration/multifd/tcp/tls/x509/override-host > # /x86_64/migration/multifd/tcp/tls/x509/mismatch-host > # /x86_64/migration/multifd/tcp/tls/x509/allow-anon-client > # /x86_64/migration/multifd/tcp/tls/x509/reject-anon-client > # /x86_64/migration/validate_uri/channels/both_set > # /x86_64/migration/validate_uri/channels/none_set > > Individually none of those is very slow on its own - 10 are in > the 2-3 second range, 35 are 1-2 secs, and 6 are less than > 1 second. > > A very large portion of those are validating different ways to > establish migration. Hardly any of them actually need to run > a migration to completion. Even without running to completion > though, we have the overheads of spawning 2 QEMUs. > > This feels like something that should be amenable to unit testing. > Might need a little re-factoring of migration code to make it > easier to run a subset of its logic in isolation, but that'd > probably be a win anyway, as such work usually makes code cleaner. I'll invest some time in that. I've no idea how we do unit testing in QEMU. > >> Do we need nightly CI runs? Unit tests? Bear in mind there's a resource >> allocation issue there. Addressing problems with timeouts/races in our >> CI is not something any random person can do. > > In terms of running time, I think migration-test is acceptable as it is > to run in 'make check' by default and doesn't justify dropping test > coverage. We should still look to optimize & move to unit testing more > code, and any reliability issues are something that needs to be addressed > too. > > With regards, > Daniel
On Fri, Oct 18, 2024 at 10:51:03AM -0300, Fabiano Rosas wrote: > Daniel P. Berrangé <berrange@redhat.com> writes: > > > On Fri, Oct 18, 2024 at 10:46:55AM +0100, Peter Maydell wrote: > >> On Fri, 18 Oct 2024 at 10:01, Daniel P. Berrangé <berrange@redhat.com> wrote: > >> > > >> > On Thu, Oct 17, 2024 at 01:29:35PM -0300, Fabiano Rosas wrote: > >> > > Daniel P. Berrangé <berrange@redhat.com> writes: > >> > > > >> > > > On Thu, Oct 17, 2024 at 11:32:11AM -0300, Fabiano Rosas wrote: > >> > > >> Recent changes to how we invoke the migration tests have > >> > > >> (intentionally) caused them to not be part of the check-qtest target > >> > > >> anymore. Add the check-migration-quick target so we don't lose > >> > > >> migration code testing in this job. > >> > > > > >> > > > But 'check-migration-quick' is only the subset of migration tests, > >> > > > 'check-migration' is all of the migration tests. So surely this is > >> > > > a massive regressions in covage in CI pipelines. > >> > > > >> > > I'm not sure it is. There are tests there already for all the major > >> > > parts of the code: precopy, postcopy, multifd, socket. Besides, we can > >> > > tweak migration-quick to cover spots where we think we're losing > >> > > coverage. > >> > > >> > Each of the tests in migration-test were added for a good reason, > >> > generally to address testing gaps where we had functional regressions > >> > in the past. I don't think its a good idea to stop running such tests > >> > in CI as gating on new contributions. Any time we've had optional > >> > tests in QEMU, we've seen repeated regressions in the area in question. > >> > > >> > > Since our CI offers nothing in terms of reproducibility or > >> > > debuggability, I don't think it's productive to have an increasing > >> > > amount of tests running in CI if that means we'll be dealing with > >> > > timeouts and intermittent crashes constantly. > >> > > >> > Test reliability is a different thing. If a particular test is > >> > flaky, it needs to either be fixed or disabled. Splitting into > >> > a fast & slow grouping doesn't address reliability, just hides > >> > the problem from view. > >> > >> A lot of the current reliability issue is timeouts -- sometimes > >> our CI runners just run really slow (I have seen an example where > >> between a normal and a slow run on the same commit both the > >> compile and test times were 10x different...) So any test > >> that is not a fast-to-complete is much much more likely to > >> hit its timeout if the runner is running slowly. When I am > >> doing CI testing for merges "migration test timed out again" > >> is really really common. > > > > If its frequently timing out, then we've got the timeouts > > wrong, or we have some genuine bugs in there to be fixed. > > > >> > > No disagreement here. But then I'm going to need advice on what to do > >> > > when other maintainers ask us to stop writing migration tests because > >> > > they take too long. I cannot send contributors away nor merge code > >> > > without tests. > >> > > >> > In general, I think it is unreasonable for other maintainers to > >> > tell us to stop adding test coverage for migration, and would > >> > push back against such a request. > >> > >> We do not have infinite CI resources, unfortunately. Migration > >> is competing with everything else for time on CI. You have to > >> find a balance between "what do we run every time" and "what > >> do we only run when specifically testing a migration pullreq". > >> Similarly, there's a lot of iotests but we don't run all of them > >> for every block backend for every CI job via "make check". > > > > The combos we don't run for iotests are a good source of > > regressions too :-( > > > >> Long test times for tests run under "make check" are also bad > >> for individual developers -- if I'm running "make check" to > >> test a target/arm change I've made I don't really want that > >> to then spend 15 minutes testing the migration code that > >> I haven't touched and that is vanishingly unlikely to be > >> affected by my patches. > > > > Migration-test *used* to take 15 minutes to run, but that was a > > very long time ago. A run of it today is around 1m20. > > > > That said, if you are building multiple system emulators, we > > run the same test multiple times, and with the number of > > targets we have, that will be painful. > > > > That could be a good reason to split the migration-test into > > two distinct programs. One program that runs for every target, > > and one that is only run once, for some arbitrary "primary" > > target ? > > What do you mean by distinct programs? It's not the migration-test that > decides on which targets it runs, it's meson.build. We register a test() > for each target, same as with any other qtest. Maybe I misunderstood > you... If we split, we could have meson.build register "migration-smoketest" for every target while registering "migration-bigtest" for just 1 target. > > Or could we make use of glib's g_test_thorough > > for this - a primary target runs with "SPEED=through" and > > all other targets with normal settings. That would give us > > a way to optimize any of the qtests to reduce redundant > > testing where appropriate. > > This still requires a new make target I think. Otherwise we'd run *all* > thorough tests for a QEMU target and not only migration-test in thorough > mode. Yes, that's true, having separate programs is probably an easier option than playing games with "SPEED" settings. With regards, Daniel
On Fri, Oct 18, 2024 at 10:51:08AM -0300, Fabiano Rosas wrote: > Daniel P. Berrangé <berrange@redhat.com> writes: > > > On Thu, Oct 17, 2024 at 01:29:35PM -0300, Fabiano Rosas wrote: > >> Daniel P. Berrangé <berrange@redhat.com> writes: > >> > >> > On Thu, Oct 17, 2024 at 11:32:11AM -0300, Fabiano Rosas wrote: > >> >> Recent changes to how we invoke the migration tests have > >> >> (intentionally) caused them to not be part of the check-qtest target > >> >> anymore. Add the check-migration-quick target so we don't lose > >> >> migration code testing in this job. > >> > > >> > But 'check-migration-quick' is only the subset of migration tests, > >> > 'check-migration' is all of the migration tests. So surely this is > >> > a massive regressions in covage in CI pipelines. > >> > >> I'm not sure it is. There are tests there already for all the major > >> parts of the code: precopy, postcopy, multifd, socket. Besides, we can > >> tweak migration-quick to cover spots where we think we're losing > >> coverage. > > > > Each of the tests in migration-test were added for a good reason, > > generally to address testing gaps where we had functional regressions > > in the past. I don't think its a good idea to stop running such tests > > in CI as gating on new contributions. Any time we've had optional > > tests in QEMU, we've seen repeated regressions in the area in question. > > > >> Since our CI offers nothing in terms of reproducibility or > >> debuggability, I don't think it's productive to have an increasing > >> amount of tests running in CI if that means we'll be dealing with > >> timeouts and intermittent crashes constantly. > > > > Test reliability is a different thing. If a particular test is > > flaky, it needs to either be fixed or disabled. > > The problem is that in this community the idea of "fix" is: wait until > someone with the appropriate skill level and interest stumbles upon the > problem on their own and fix it in anger. > > For it to be a proper strategy, we'd need to create an issue in gitlab > referencing the bug, have a proper reproducer and encourage contributors > to work on the issue. It is policy that we should be creating issues in gitlab for any flaky tests. I wouldn't say we've been perfect at that, but we should be doing it, and that link ought to be linked in the code if we disable the test there. > - It was disabled in March 2023 and stood there *not testing anything* > while a major refactoring of the test code was happening. > > - The test was fixed in June 2023, but not reenabled in fear of getting > flak from the community for breaking CI again (or at least that's the > feeling I got from talking to Juan). > > - mapped-ram (which relies entirely on multifd) started being worked on > and I had to enable the test in my own branch to be able to test the > code properly. While disabled, it caught several issues in mapped-ram. > > - In October 2023 the test is re-enabled an immediately exposes issues > in the code. > > This is how I started working on the migration code. Maybe you can > appreciate why I don't feel confident about this fix or disable > strategy. It has eaten many hours of my work. The migration subsystem was definitely suffering from insufficient maintainer resources available historically, which is reflected in some of the testing problems we've had & largely how I ended up spending so much time on migration code too. > > Looking at its execution time right now, I'd say migration test > > is pretty good, considering the permutations we have to target. > > > > It gets a bad reputation because historically it has been as > > much as x20 slower than it is today, and has also struggled > > with reliability. The latter is a reflection of the complexity > > of migration and and IMHO actually justifies greater testing, > > as long as we put in time to address bugs. > > > > Also we've got one single test program, covering an entire > > subsystem in one go, rather than lots of short individual > > test programs, so migration unfairly gets blamed for being > > slow, when it simply covers alot of functionality in one > > program. > > And still you think it's not worth it having a separate target for > testing migration. FWIW, I also proposed splittling it into multiple > meson tests, which you also rejected. It would be so much easier to move > all of this into a separate target and let those who want nothing do to > with to just ignore it. In the qtests/meson.build, I see we register separate suites - a generic "qtest" suite, and a "qtest-$TARGET" suite. What's missing here is a suite for subsystem classification, which I guess is more or less what you proposed here for migration. How about (in addition to the idea of splitting migration-test into one part run for all targets, and one part run for just one target), we generalize this concept to work for any subsystem tagging qtest_subsystems = { 'migration-test': ['migration'], 'cdrom-test': ['block'], 'ahci-test': ['block'], .... } then when registering tests we could do suites = ['qtest', 'qtest-' + target_base] foreach subsys: qtest_subsystems.get(test, []) suites += ['qtest-' + subsys, 'qtest-' + target_base + '-' + subsys] endforeach test(.... suite: suites) that would give us a way to run just the migration tests, or just the migration tests on x86_64, etc, likewise for other subsystems we want to tag, while still keeping 'make check-qtest' as the full coverage. > >> > Any tests in tree need to be exercised by CI as the minimum bar > >> > to prevent bit rot from merges. > >> > > >> > >> No disagreement here. But then I'm going to need advice on what to do > >> when other maintainers ask us to stop writing migration tests because > >> they take too long. I cannot send contributors away nor merge code > >> without tests. > > > > In general, I think it is unreasonable for other maintainers to > > tell us to stop adding test coverage for migration, and would > > push back against such a request. > > This might be a larger issue in QEMU. I also heard the same back in 2021 > when doing ppc work: "don't add too many tests because the CI buckles > and people get mad". The same with adding too much logging really. We're > hostages to the gitlab CI unfortunately. Yep, we need more investment in our CI resources. There were some ideas discussed at KVM Forum that could help with this, which should be publicised soon. > > This feels like something that should be amenable to unit testing. > > Might need a little re-factoring of migration code to make it > > easier to run a subset of its logic in isolation, but that'd > > probably be a win anyway, as such work usually makes code cleaner. > > I'll invest some time in that. I've no idea how we do unit testing in > QEMU. Mostly starts with the standard glib test program setup, where by you create a tests/unit/test-<blah>.c file, with functions for each test case that you register with g_test_add, same as qtest. The key difference from qtest is that you then just directly link the test binary to the code to be tested, and call into it internal QEMU APIs directly. In this case, it would involve linking to the 'libmigration' meson static library object. With regards, Daniel
Daniel P. Berrangé <berrange@redhat.com> writes: > On Fri, Oct 18, 2024 at 10:51:03AM -0300, Fabiano Rosas wrote: >> Daniel P. Berrangé <berrange@redhat.com> writes: >> >> > On Fri, Oct 18, 2024 at 10:46:55AM +0100, Peter Maydell wrote: >> >> On Fri, 18 Oct 2024 at 10:01, Daniel P. Berrangé <berrange@redhat.com> wrote: >> >> > >> >> > On Thu, Oct 17, 2024 at 01:29:35PM -0300, Fabiano Rosas wrote: >> >> > > Daniel P. Berrangé <berrange@redhat.com> writes: >> >> > > >> >> > > > On Thu, Oct 17, 2024 at 11:32:11AM -0300, Fabiano Rosas wrote: >> >> > > >> Recent changes to how we invoke the migration tests have >> >> > > >> (intentionally) caused them to not be part of the check-qtest target >> >> > > >> anymore. Add the check-migration-quick target so we don't lose >> >> > > >> migration code testing in this job. >> >> > > > >> >> > > > But 'check-migration-quick' is only the subset of migration tests, >> >> > > > 'check-migration' is all of the migration tests. So surely this is >> >> > > > a massive regressions in covage in CI pipelines. >> >> > > >> >> > > I'm not sure it is. There are tests there already for all the major >> >> > > parts of the code: precopy, postcopy, multifd, socket. Besides, we can >> >> > > tweak migration-quick to cover spots where we think we're losing >> >> > > coverage. >> >> > >> >> > Each of the tests in migration-test were added for a good reason, >> >> > generally to address testing gaps where we had functional regressions >> >> > in the past. I don't think its a good idea to stop running such tests >> >> > in CI as gating on new contributions. Any time we've had optional >> >> > tests in QEMU, we've seen repeated regressions in the area in question. >> >> > >> >> > > Since our CI offers nothing in terms of reproducibility or >> >> > > debuggability, I don't think it's productive to have an increasing >> >> > > amount of tests running in CI if that means we'll be dealing with >> >> > > timeouts and intermittent crashes constantly. >> >> > >> >> > Test reliability is a different thing. If a particular test is >> >> > flaky, it needs to either be fixed or disabled. Splitting into >> >> > a fast & slow grouping doesn't address reliability, just hides >> >> > the problem from view. >> >> >> >> A lot of the current reliability issue is timeouts -- sometimes >> >> our CI runners just run really slow (I have seen an example where >> >> between a normal and a slow run on the same commit both the >> >> compile and test times were 10x different...) So any test >> >> that is not a fast-to-complete is much much more likely to >> >> hit its timeout if the runner is running slowly. When I am >> >> doing CI testing for merges "migration test timed out again" >> >> is really really common. >> > >> > If its frequently timing out, then we've got the timeouts >> > wrong, or we have some genuine bugs in there to be fixed. >> > >> >> > > No disagreement here. But then I'm going to need advice on what to do >> >> > > when other maintainers ask us to stop writing migration tests because >> >> > > they take too long. I cannot send contributors away nor merge code >> >> > > without tests. >> >> > >> >> > In general, I think it is unreasonable for other maintainers to >> >> > tell us to stop adding test coverage for migration, and would >> >> > push back against such a request. >> >> >> >> We do not have infinite CI resources, unfortunately. Migration >> >> is competing with everything else for time on CI. You have to >> >> find a balance between "what do we run every time" and "what >> >> do we only run when specifically testing a migration pullreq". >> >> Similarly, there's a lot of iotests but we don't run all of them >> >> for every block backend for every CI job via "make check". >> > >> > The combos we don't run for iotests are a good source of >> > regressions too :-( >> > >> >> Long test times for tests run under "make check" are also bad >> >> for individual developers -- if I'm running "make check" to >> >> test a target/arm change I've made I don't really want that >> >> to then spend 15 minutes testing the migration code that >> >> I haven't touched and that is vanishingly unlikely to be >> >> affected by my patches. >> > >> > Migration-test *used* to take 15 minutes to run, but that was a >> > very long time ago. A run of it today is around 1m20. >> > >> > That said, if you are building multiple system emulators, we >> > run the same test multiple times, and with the number of >> > targets we have, that will be painful. >> > >> > That could be a good reason to split the migration-test into >> > two distinct programs. One program that runs for every target, >> > and one that is only run once, for some arbitrary "primary" >> > target ? >> >> What do you mean by distinct programs? It's not the migration-test that >> decides on which targets it runs, it's meson.build. We register a test() >> for each target, same as with any other qtest. Maybe I misunderstood >> you... > > If we split, we could have meson.build register "migration-smoketest" > for every target while registering "migration-bigtest" for just 1 target. Isn't that a bunch of shuffling code around just to have two different invocations of migration-test? There's the possibility of using the gtester path like /x86_64/migration/smoke and passing '-r' when invoking via meson. That requires way fewer changes in the C code. It moves the complexity into meson.build, which might be worse. Alex's idea of full set for KVM arch + a TCG arch would probably be trickier in meson. > >> > Or could we make use of glib's g_test_thorough >> > for this - a primary target runs with "SPEED=through" and >> > all other targets with normal settings. That would give us >> > a way to optimize any of the qtests to reduce redundant >> > testing where appropriate. >> >> This still requires a new make target I think. Otherwise we'd run *all* >> thorough tests for a QEMU target and not only migration-test in thorough >> mode. > > Yes, that's true, having separate programs is probably an easier > option than playing games with "SPEED" settings. > > > With regards, > Daniel
On Fri, Oct 18, 2024 at 11:28:53AM -0300, Fabiano Rosas wrote: > Daniel P. Berrangé <berrange@redhat.com> writes: > > > On Fri, Oct 18, 2024 at 10:51:03AM -0300, Fabiano Rosas wrote: > >> Daniel P. Berrangé <berrange@redhat.com> writes: > >> > >> > That could be a good reason to split the migration-test into > >> > two distinct programs. One program that runs for every target, > >> > and one that is only run once, for some arbitrary "primary" > >> > target ? > >> > >> What do you mean by distinct programs? It's not the migration-test that > >> decides on which targets it runs, it's meson.build. We register a test() > >> for each target, same as with any other qtest. Maybe I misunderstood > >> you... > > > > If we split, we could have meson.build register "migration-smoketest" > > for every target while registering "migration-bigtest" for just 1 target. > > Isn't that a bunch of shuffling code around just to have two different > invocations of migration-test? Yes, pretty much, but that's not an inherantly bad thing. Migration is a bit of an outlier in its attempt to test the entire subsystem from a single test binary. > There's the possibility of using the gtester path like > /x86_64/migration/smoke and passing '-r' when invoking via meson. That > requires way fewer changes in the C code. It moves the complexity into > meson.build, which might be worse. > > Alex's idea of full set for KVM arch + a TCG arch would probably be > trickier in meson. With regards, Daniel
Daniel P. Berrangé <berrange@redhat.com> writes: > On Fri, Oct 18, 2024 at 10:51:08AM -0300, Fabiano Rosas wrote: >> Daniel P. Berrangé <berrange@redhat.com> writes: >> >> > On Thu, Oct 17, 2024 at 01:29:35PM -0300, Fabiano Rosas wrote: >> >> Daniel P. Berrangé <berrange@redhat.com> writes: >> >> >> >> > On Thu, Oct 17, 2024 at 11:32:11AM -0300, Fabiano Rosas wrote: >> >> >> Recent changes to how we invoke the migration tests have >> >> >> (intentionally) caused them to not be part of the check-qtest target >> >> >> anymore. Add the check-migration-quick target so we don't lose >> >> >> migration code testing in this job. >> >> > >> >> > But 'check-migration-quick' is only the subset of migration tests, >> >> > 'check-migration' is all of the migration tests. So surely this is >> >> > a massive regressions in covage in CI pipelines. >> >> >> >> I'm not sure it is. There are tests there already for all the major >> >> parts of the code: precopy, postcopy, multifd, socket. Besides, we can >> >> tweak migration-quick to cover spots where we think we're losing >> >> coverage. >> > >> > Each of the tests in migration-test were added for a good reason, >> > generally to address testing gaps where we had functional regressions >> > in the past. I don't think its a good idea to stop running such tests >> > in CI as gating on new contributions. Any time we've had optional >> > tests in QEMU, we've seen repeated regressions in the area in question. >> > >> >> Since our CI offers nothing in terms of reproducibility or >> >> debuggability, I don't think it's productive to have an increasing >> >> amount of tests running in CI if that means we'll be dealing with >> >> timeouts and intermittent crashes constantly. >> > >> > Test reliability is a different thing. If a particular test is >> > flaky, it needs to either be fixed or disabled. >> >> The problem is that in this community the idea of "fix" is: wait until >> someone with the appropriate skill level and interest stumbles upon the >> problem on their own and fix it in anger. >> >> For it to be a proper strategy, we'd need to create an issue in gitlab >> referencing the bug, have a proper reproducer and encourage contributors >> to work on the issue. > > It is policy that we should be creating issues in gitlab for any > flaky tests. I wouldn't say we've been perfect at that, but we > should be doing it, and that link ought to be linked in the code > if we disable the test there. > >> - It was disabled in March 2023 and stood there *not testing anything* >> while a major refactoring of the test code was happening. >> >> - The test was fixed in June 2023, but not reenabled in fear of getting >> flak from the community for breaking CI again (or at least that's the >> feeling I got from talking to Juan). >> >> - mapped-ram (which relies entirely on multifd) started being worked on >> and I had to enable the test in my own branch to be able to test the >> code properly. While disabled, it caught several issues in mapped-ram. >> >> - In October 2023 the test is re-enabled an immediately exposes issues >> in the code. >> >> This is how I started working on the migration code. Maybe you can >> appreciate why I don't feel confident about this fix or disable >> strategy. It has eaten many hours of my work. > > The migration subsystem was definitely suffering from insufficient > maintainer resources available historically, which is reflected > in some of the testing problems we've had & largely how I ended > up spending so much time on migration code too. > >> > Looking at its execution time right now, I'd say migration test >> > is pretty good, considering the permutations we have to target. >> > >> > It gets a bad reputation because historically it has been as >> > much as x20 slower than it is today, and has also struggled >> > with reliability. The latter is a reflection of the complexity >> > of migration and and IMHO actually justifies greater testing, >> > as long as we put in time to address bugs. >> > >> > Also we've got one single test program, covering an entire >> > subsystem in one go, rather than lots of short individual >> > test programs, so migration unfairly gets blamed for being >> > slow, when it simply covers alot of functionality in one >> > program. >> >> And still you think it's not worth it having a separate target for >> testing migration. FWIW, I also proposed splittling it into multiple >> meson tests, which you also rejected. It would be so much easier to move >> all of this into a separate target and let those who want nothing do to >> with to just ignore it. > > In the qtests/meson.build, I see we register separate > suites - a generic "qtest" suite, and a "qtest-$TARGET" > suite. What's missing here is a suite for subsystem > classification, which I guess is more or less what you > proposed here for migration. > > How about (in addition to the idea of splitting migration-test > into one part run for all targets, and one part run for just > one target), we generalize this concept to work for any > subsystem tagging > > qtest_subsystems = { > 'migration-test': ['migration'], > 'cdrom-test': ['block'], > 'ahci-test': ['block'], > .... > } This is interesting because it would allow a test to be considered in more than one subsystem. There are many tests that invoke migration themselves. > > > then when registering tests we could do > > suites = ['qtest', 'qtest-' + target_base] > foreach subsys: qtest_subsystems.get(test, []) > suites += ['qtest-' + subsys, 'qtest-' + target_base + '-' + subsys] Hopefully meson won't take this as a hint to construct a 1000 characters long line when invoking the test. > endforeach > > test(.... > suite: suites) > > that would give us a way to run just the migration tests, or > just the migration tests on x86_64, etc, likewise for other > subsystems we want to tag, while still keeping 'make check-qtest' > as the full coverage. > >> >> > Any tests in tree need to be exercised by CI as the minimum bar >> >> > to prevent bit rot from merges. >> >> > >> >> >> >> No disagreement here. But then I'm going to need advice on what to do >> >> when other maintainers ask us to stop writing migration tests because >> >> they take too long. I cannot send contributors away nor merge code >> >> without tests. >> > >> > In general, I think it is unreasonable for other maintainers to >> > tell us to stop adding test coverage for migration, and would >> > push back against such a request. >> >> This might be a larger issue in QEMU. I also heard the same back in 2021 >> when doing ppc work: "don't add too many tests because the CI buckles >> and people get mad". The same with adding too much logging really. We're >> hostages to the gitlab CI unfortunately. > > Yep, we need more investment in our CI resources. There were some > ideas discussed at KVM Forum that could help with this, which > should be publicised soon. Great. Looking forward to those. > > >> > This feels like something that should be amenable to unit testing. >> > Might need a little re-factoring of migration code to make it >> > easier to run a subset of its logic in isolation, but that'd >> > probably be a win anyway, as such work usually makes code cleaner. >> >> I'll invest some time in that. I've no idea how we do unit testing in >> QEMU. > > Mostly starts with the standard glib test program setup, where by > you create a tests/unit/test-<blah>.c file, with functions for > each test case that you register with g_test_add, same as qtest. > > The key difference from qtest is that you then just directly > link the test binary to the code to be tested, and call into > it internal QEMU APIs directly. In this case, it would involve > linking to the 'libmigration' meson static library object. Thanks for the introduction. I'll pick some self-contained part of migration and see how far we are from being able to write unit tests. > > > With regards, > Daniel
On Fri, 18 Oct 2024 at 10:01, Daniel P. Berrangé <berrange@redhat.com> wrote: > Test reliability is a different thing. If a particular test is > flaky, it needs to either be fixed or disabled. Splitting into > a fast & slow grouping doesn't address reliability, just hides > the problem from view. On the subject of 'flaky', here's another low-repeatability intermittent with migration-test that I just ran into in 'make vm-build-openbsd': ▶ 97/916 /ppc64/migration/multifd/tcp/plain/cancel OK ▶ 96/916 /i386/migration/precopy/tcp/tls/x509/allow-anon-client OK ▶ 97/916 /ppc64/migration/multifd/tcp/plain/zlib - ERROR:../src/tests/qtest/migration-helpers.c:322:check_migration_status: assertion failed (current_status != "failed"): ("failed" != "failed") FAIL ▶ 97/916 ERROR ▶ 95/916 /aarch64/migration/multifd/tcp/channels/plain/none OK 97/916 qemu:qtest+qtest-ppc64 / qtest-ppc64/migration-test ERROR 134.38s killed by signal 6 SIGABRT ――――――――――――――――――――――――――――――――――――― ✀ ――――――――――――――――――――――――――――――――――――― stderr: warning: fd: migration to a file is deprecated. Use file: instead. warning: fd: migration to a file is deprecated. Use file: instead. ** ERROR:../src/tests/qtest/migration-helpers.c:322:check_migration_status: assertion failed (current_status != "failed"): ("failed" != "failed") qemu-system-ppc64: Failed to connect to '127.0.0.1:24109': Address already in use qemu-system-ppc64: Failed to peek at channel (test program exited with status code -6) ―――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――― Probably https://lore.kernel.org/qemu-devel/CAFEAcA8p9BKCrn9EfFXTpHE+5w-_8zhtE_52SpZLuS-+zpF5Gg@mail.gmail.com/ in a slightly different form. thanks -- PMM
On Fri, Oct 18, 2024 at 04:25:07PM +0100, Peter Maydell wrote: > On Fri, 18 Oct 2024 at 10:01, Daniel P. Berrangé <berrange@redhat.com> wrote: > > Test reliability is a different thing. If a particular test is > > flaky, it needs to either be fixed or disabled. Splitting into > > a fast & slow grouping doesn't address reliability, just hides > > the problem from view. > > On the subject of 'flaky', here's another low-repeatability > intermittent with migration-test that I just ran into in > 'make vm-build-openbsd': > > ▶ 97/916 /ppc64/migration/multifd/tcp/plain/cancel > OK > ▶ 96/916 /i386/migration/precopy/tcp/tls/x509/allow-anon-client > OK > ▶ 97/916 /ppc64/migration/multifd/tcp/plain/zlib - > ERROR:../src/tests/qtest/migration-helpers.c:322:check_migration_status: > assertion failed (current_status != "failed"): ("failed" != "failed") > FAIL > ▶ 97/916 > ERROR > ▶ 95/916 /aarch64/migration/multifd/tcp/channels/plain/none > OK > 97/916 qemu:qtest+qtest-ppc64 / qtest-ppc64/migration-test > ERROR 134.38s killed by signal 6 SIGABRT > ――――――――――――――――――――――――――――――――――――― ✀ ――――――――――――――――――――――――――――――――――――― > stderr: > warning: fd: migration to a file is deprecated. Use file: instead. > warning: fd: migration to a file is deprecated. Use file: instead. > ** > ERROR:../src/tests/qtest/migration-helpers.c:322:check_migration_status: > assertion failed (current_status != "failed"): ("failed" != "failed") > qemu-system-ppc64: Failed to connect to '127.0.0.1:24109': Address > already in use This is interesting, as I suspect it is a sign of a genuine portability problem, as there are some subtle sockets binding differences between Linux and *BSDs IIRC. With regards, Daniel
On Fri, Oct 18, 2024 at 04:25:07PM +0100, Peter Maydell wrote: > On Fri, 18 Oct 2024 at 10:01, Daniel P. Berrangé <berrange@redhat.com> wrote: > > Test reliability is a different thing. If a particular test is > > flaky, it needs to either be fixed or disabled. Splitting into > > a fast & slow grouping doesn't address reliability, just hides > > the problem from view. > > On the subject of 'flaky', here's another low-repeatability > intermittent with migration-test that I just ran into in > 'make vm-build-openbsd': > ――――――――――――――――――――――――――――――――――――― ✀ ――――――――――――――――――――――――――――――――――――― > stderr: > warning: fd: migration to a file is deprecated. Use file: instead. > warning: fd: migration to a file is deprecated. Use file: instead. > ** > ERROR:../src/tests/qtest/migration-helpers.c:322:check_migration_status: > assertion failed (current_status != "failed"): ("failed" != "failed") > qemu-system-ppc64: Failed to connect to '127.0.0.1:24109': Address > already in use > qemu-system-ppc64: Failed to peek at channel > > (test program exited with status code -6) > ―――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――― > > Probably https://lore.kernel.org/qemu-devel/CAFEAcA8p9BKCrn9EfFXTpHE+5w-_8zhtE_52SpZLuS-+zpF5Gg@mail.gmail.com/ I think I've finally found the root cause of this bug. Setting SO_REUSEADDR on client sockets on OpenBSD causes re-use of ports in TIME_WAIT when auto-assigning a local bind address for connections. I've sent a patch to remove this, since it is essentially pointless todo this AFAIK. With regards, Daniel
diff --git a/.gitlab-ci.d/buildtest.yml b/.gitlab-ci.d/buildtest.yml index 34d3f4e9ab..37247dc5fa 100644 --- a/.gitlab-ci.d/buildtest.yml +++ b/.gitlab-ci.d/buildtest.yml @@ -442,7 +442,7 @@ clang-system: CONFIGURE_ARGS: --cc=clang --cxx=clang++ --enable-ubsan --extra-cflags=-fno-sanitize-recover=undefined TARGETS: alpha-softmmu arm-softmmu m68k-softmmu mips64-softmmu s390x-softmmu - MAKE_CHECK_ARGS: check-qtest check-tcg + MAKE_CHECK_ARGS: check-qtest check-tcg check-migration-quick clang-user: extends: .native_build_job_template
Recent changes to how we invoke the migration tests have (intentionally) caused them to not be part of the check-qtest target anymore. Add the check-migration-quick target so we don't lose migration code testing in this job. Signed-off-by: Fabiano Rosas <farosas@suse.de> --- .gitlab-ci.d/buildtest.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-)