Message ID | cover.1667224593.git.bertrand.marquis@arm.com (mailing list archive) |
---|---|
Headers | show |
Series | Yocto Gitlab CI | expand |
Hi Bertrand, On 31/10/2022 15:00, Bertrand Marquis wrote: > > > This patch series is a first attempt to check if we could use Yocto in > gitlab ci to build and run xen on qemu for arm, arm64 and x86. > > The first patch is creating a container with all elements required to > build Yocto, a checkout of the yocto layers required and an helper > script to build and run xen on qemu with yocto. > > The second patch is creating containers with a first build of yocto done > so that susbsequent build with those containers would only rebuild what > was changed and take the rest from the cache. > > The third patch is adding a way to easily clean locally created > containers. > > This is is mainly for discussion and sharing as there are still some > issues/problem to solve: > - building the qemu* containers can take several hours depending on the > network bandwith and computing power of the machine where those are > created This is not really an issue as the build of the containers occurs on the local machines before pushing them to registry. Also, building the containers will only be required for new Yocto releases. > - produced containers containing the cache have a size between 8 and > 12GB depending on the architecture. We might need to store the build > cache somewhere else to reduce the size. If we choose to have one > single image, the needed size is around 20GB and we need up to 40GB > during the build, which is why I splitted them. > - during the build and run, we use a bit more then 20GB of disk which is > over the allowed size in gitlab As we could see during v2 testing, we do not have any space restrictions on the Xen GitLab and I think we already decided to have the Yocto integrated into our CI. I will do some testing and get back to you with results + review. ~Michal
Hi Michal, > On 31 Oct 2022, at 14:39, Michal Orzel <michal.orzel@amd.com> wrote: > > Hi Bertrand, > > On 31/10/2022 15:00, Bertrand Marquis wrote: >> >> >> This patch series is a first attempt to check if we could use Yocto in >> gitlab ci to build and run xen on qemu for arm, arm64 and x86. >> >> The first patch is creating a container with all elements required to >> build Yocto, a checkout of the yocto layers required and an helper >> script to build and run xen on qemu with yocto. >> >> The second patch is creating containers with a first build of yocto done >> so that susbsequent build with those containers would only rebuild what >> was changed and take the rest from the cache. >> >> The third patch is adding a way to easily clean locally created >> containers. >> >> This is is mainly for discussion and sharing as there are still some >> issues/problem to solve: >> - building the qemu* containers can take several hours depending on the >> network bandwith and computing power of the machine where those are >> created > This is not really an issue as the build of the containers occurs on the local > machines before pushing them to registry. Also, building the containers > will only be required for new Yocto releases. > >> - produced containers containing the cache have a size between 8 and >> 12GB depending on the architecture. We might need to store the build >> cache somewhere else to reduce the size. If we choose to have one >> single image, the needed size is around 20GB and we need up to 40GB >> during the build, which is why I splitted them. >> - during the build and run, we use a bit more then 20GB of disk which is >> over the allowed size in gitlab > As we could see during v2 testing, we do not have any space restrictions > on the Xen GitLab and I think we already decided to have the Yocto > integrated into our CI. Right, I should have modified this chapter to be coherent with your latest tests. Sorry for that. > > I will do some testing and get back to you with results + review. Thanks Cheers Bertrand > > ~Michal
Hi Bertrand and Stefano, On 31/10/2022 16:00, Bertrand Marquis wrote: > > > Hi Michal, > >> On 31 Oct 2022, at 14:39, Michal Orzel <michal.orzel@amd.com> wrote: >> >> Hi Bertrand, >> >> On 31/10/2022 15:00, Bertrand Marquis wrote: >>> >>> >>> This patch series is a first attempt to check if we could use Yocto in >>> gitlab ci to build and run xen on qemu for arm, arm64 and x86. >>> >>> The first patch is creating a container with all elements required to >>> build Yocto, a checkout of the yocto layers required and an helper >>> script to build and run xen on qemu with yocto. >>> >>> The second patch is creating containers with a first build of yocto done >>> so that susbsequent build with those containers would only rebuild what >>> was changed and take the rest from the cache. >>> >>> The third patch is adding a way to easily clean locally created >>> containers. >>> >>> This is is mainly for discussion and sharing as there are still some >>> issues/problem to solve: >>> - building the qemu* containers can take several hours depending on the >>> network bandwith and computing power of the machine where those are >>> created >> This is not really an issue as the build of the containers occurs on the local >> machines before pushing them to registry. Also, building the containers >> will only be required for new Yocto releases. >> >>> - produced containers containing the cache have a size between 8 and >>> 12GB depending on the architecture. We might need to store the build >>> cache somewhere else to reduce the size. If we choose to have one >>> single image, the needed size is around 20GB and we need up to 40GB >>> during the build, which is why I splitted them. >>> - during the build and run, we use a bit more then 20GB of disk which is >>> over the allowed size in gitlab >> As we could see during v2 testing, we do not have any space restrictions >> on the Xen GitLab and I think we already decided to have the Yocto >> integrated into our CI. > > Right, I should have modified this chapter to be coherent with your latest tests. > Sorry for that. > >> >> I will do some testing and get back to you with results + review. I did some testing and here are the results: In the current form this series will fail when running CI because the Yocto containers are based on "From ubuntu:22.04" (there is no platform prefix), which means that the containers are built for the host architecture (in my case and in 99% of the cases of the local build it will be x86). In Gitlab we have 2 runners (arm64 and x86_64). This means that all the test jobs would need to specify x86_64 as a tag when keeping the current behavior. After I built all the containers on my x86 machine, I pushed them to registry and the pipeline was successful: https://gitlab.com/xen-project/people/morzel/xen-orzelmichal/-/pipelines/686853939 Here is the diff on patch no. 3 to make the series work (using x86 tag and small improvement to include needs: []): ``` diff --git a/automation/gitlab-ci/test.yaml b/automation/gitlab-ci/test.yaml index 5c620fefce59..52cccec6f904 100644 --- a/automation/gitlab-ci/test.yaml +++ b/automation/gitlab-ci/test.yaml @@ -65,6 +65,9 @@ paths: - 'logs/*' when: always + needs: [] + tags: + - x86_64 # Test jobs build-each-commit-gcc: @@ -206,19 +209,13 @@ yocto-qemuarm64: extends: .yocto-test variables: YOCTO_BOARD: qemuarm64 - tags: - - arm64 yocto-qemuarm: extends: .yocto-test variables: YOCTO_BOARD: qemuarm - tags: - - arm32 yocto-qemux86-64: extends: .yocto-test variables: YOCTO_BOARD: qemux86-64 - tags: - - x86_64 ``` Now, the logical way would be to build x86 yocto container for x86, arm64 for arm64 and arm32 on arm64 or x86. I tried building the container qemuarm64 specifying target arm64 on x86. After 15h, only 70% of the Yocto build was completed and there was an error with glibc (the local build of the container for the host arch takes on my machine max 2h). This enormous amount of time is due to the qemu docker emulation that happens behind the scenes (I checked on 2 different machines). So we have 3 solutions: 1) Build and run these containers for/on x86_64: - local users can build the containers on local machines that are almost always x86 based, in short period of time, - "everyone" can build/push the containers once there is a new Yocto release - slightly slower CI build time 2) Build and run these containers for specific architectures: - almost no go for local users using x86 machine (unless using more than 16 threads (which I used) and willing to wait 2 days for the build) - faster CI build time (arm64 runner is faster than x86 one) - someone with arm64 based machine (not that common) would have to build and push the containers 3) Try to use CI to build and push the containers to registry - it could be possible but what about local users Regardless of what we chose, we need to keep in mind that the biggest advantage to the Yocto build/run is that it allows/should allow local users to perform basic testing for all the Xen supported architectures. This is because everything happens in one place with one command. ~Michal
Hi Michal, > On 7 Nov 2022, at 07:50, Michal Orzel <michal.orzel@amd.com> wrote: > > Hi Bertrand and Stefano, > > On 31/10/2022 16:00, Bertrand Marquis wrote: >> >> >> Hi Michal, >> >>> On 31 Oct 2022, at 14:39, Michal Orzel <michal.orzel@amd.com> wrote: >>> >>> Hi Bertrand, >>> >>> On 31/10/2022 15:00, Bertrand Marquis wrote: >>>> >>>> >>>> This patch series is a first attempt to check if we could use Yocto in >>>> gitlab ci to build and run xen on qemu for arm, arm64 and x86. >>>> >>>> The first patch is creating a container with all elements required to >>>> build Yocto, a checkout of the yocto layers required and an helper >>>> script to build and run xen on qemu with yocto. >>>> >>>> The second patch is creating containers with a first build of yocto done >>>> so that susbsequent build with those containers would only rebuild what >>>> was changed and take the rest from the cache. >>>> >>>> The third patch is adding a way to easily clean locally created >>>> containers. >>>> >>>> This is is mainly for discussion and sharing as there are still some >>>> issues/problem to solve: >>>> - building the qemu* containers can take several hours depending on the >>>> network bandwith and computing power of the machine where those are >>>> created >>> This is not really an issue as the build of the containers occurs on the local >>> machines before pushing them to registry. Also, building the containers >>> will only be required for new Yocto releases. >>> >>>> - produced containers containing the cache have a size between 8 and >>>> 12GB depending on the architecture. We might need to store the build >>>> cache somewhere else to reduce the size. If we choose to have one >>>> single image, the needed size is around 20GB and we need up to 40GB >>>> during the build, which is why I splitted them. >>>> - during the build and run, we use a bit more then 20GB of disk which is >>>> over the allowed size in gitlab >>> As we could see during v2 testing, we do not have any space restrictions >>> on the Xen GitLab and I think we already decided to have the Yocto >>> integrated into our CI. >> >> Right, I should have modified this chapter to be coherent with your latest tests. >> Sorry for that. >> >>> >>> I will do some testing and get back to you with results + review. > I did some testing and here are the results: > > In the current form this series will fail when running CI because the Yocto containers > are based on "From ubuntu:22.04" (there is no platform prefix), which means that the containers > are built for the host architecture (in my case and in 99% of the cases of the local build it will > be x86). In Gitlab we have 2 runners (arm64 and x86_64). This means that all the test jobs would need > to specify x86_64 as a tag when keeping the current behavior. > After I built all the containers on my x86 machine, I pushed them to registry and the pipeline was successful: > https://gitlab.com/xen-project/people/morzel/xen-orzelmichal/-/pipelines/686853939 > > Here is the diff on patch no. 3 to make the series work (using x86 tag and small improvement to include needs: []): > ``` > diff --git a/automation/gitlab-ci/test.yaml b/automation/gitlab-ci/test.yaml > index 5c620fefce59..52cccec6f904 100644 > --- a/automation/gitlab-ci/test.yaml > +++ b/automation/gitlab-ci/test.yaml > @@ -65,6 +65,9 @@ > paths: > - 'logs/*' > when: always > + needs: [] > + tags: > + - x86_64 > > # Test jobs > build-each-commit-gcc: > @@ -206,19 +209,13 @@ yocto-qemuarm64: > extends: .yocto-test > variables: > YOCTO_BOARD: qemuarm64 > - tags: > - - arm64 > > yocto-qemuarm: > extends: .yocto-test > variables: > YOCTO_BOARD: qemuarm > - tags: > - - arm32 > > yocto-qemux86-64: > extends: .yocto-test > variables: > YOCTO_BOARD: qemux86-64 > - tags: > - - x86_64 > ``` > > Now, the logical way would be to build x86 yocto container for x86, arm64 for arm64 and arm32 on arm64 or x86. > I tried building the container qemuarm64 specifying target arm64 on x86. After 15h, only 70% of the Yocto build > was completed and there was an error with glibc (the local build of the container for the host arch takes on my machine max 2h). > This enormous amount of time is due to the qemu docker emulation that happens behind the scenes (I checked on 2 different machines). > > So we have 3 solutions: > 1) Build and run these containers for/on x86_64: > - local users can build the containers on local machines that are almost always x86 based, in short period of time, > - "everyone" can build/push the containers once there is a new Yocto release > - slightly slower CI build time > 2) Build and run these containers for specific architectures: > - almost no go for local users using x86 machine (unless using more than 16 threads (which I used) and willing to wait 2 days for the build) > - faster CI build time (arm64 runner is faster than x86 one) > - someone with arm64 based machine (not that common) would have to build and push the containers > 3) Try to use CI to build and push the containers to registry > - it could be possible but what about local users I might have a solution 4. The containers we pushed to gitlab are publicly accessible ? Shouldn’t we push both versions there and then document how someone can retrieve and use those containers instead of building them ? This way we will do the push but a normal user would just retrieve the containers and use them so building time would have less impact and we could build both flavours of the containers. > > Regardless of what we chose, we need to keep in mind that the biggest advantage to the Yocto build/run is that > it allows/should allow local users to perform basic testing for all the Xen supported architectures. This is because > everything happens in one place with one command. Agree. Bertrand > > ~Michal
Hi Bertrand, On 08/11/2022 09:42, Bertrand Marquis wrote: > > > Hi Michal, > >> On 7 Nov 2022, at 07:50, Michal Orzel <michal.orzel@amd.com> wrote: >> >> Hi Bertrand and Stefano, >> >> On 31/10/2022 16:00, Bertrand Marquis wrote: >>> >>> >>> Hi Michal, >>> >>>> On 31 Oct 2022, at 14:39, Michal Orzel <michal.orzel@amd.com> wrote: >>>> >>>> Hi Bertrand, >>>> >>>> On 31/10/2022 15:00, Bertrand Marquis wrote: >>>>> >>>>> >>>>> This patch series is a first attempt to check if we could use Yocto in >>>>> gitlab ci to build and run xen on qemu for arm, arm64 and x86. >>>>> >>>>> The first patch is creating a container with all elements required to >>>>> build Yocto, a checkout of the yocto layers required and an helper >>>>> script to build and run xen on qemu with yocto. >>>>> >>>>> The second patch is creating containers with a first build of yocto done >>>>> so that susbsequent build with those containers would only rebuild what >>>>> was changed and take the rest from the cache. >>>>> >>>>> The third patch is adding a way to easily clean locally created >>>>> containers. >>>>> >>>>> This is is mainly for discussion and sharing as there are still some >>>>> issues/problem to solve: >>>>> - building the qemu* containers can take several hours depending on the >>>>> network bandwith and computing power of the machine where those are >>>>> created >>>> This is not really an issue as the build of the containers occurs on the local >>>> machines before pushing them to registry. Also, building the containers >>>> will only be required for new Yocto releases. >>>> >>>>> - produced containers containing the cache have a size between 8 and >>>>> 12GB depending on the architecture. We might need to store the build >>>>> cache somewhere else to reduce the size. If we choose to have one >>>>> single image, the needed size is around 20GB and we need up to 40GB >>>>> during the build, which is why I splitted them. >>>>> - during the build and run, we use a bit more then 20GB of disk which is >>>>> over the allowed size in gitlab >>>> As we could see during v2 testing, we do not have any space restrictions >>>> on the Xen GitLab and I think we already decided to have the Yocto >>>> integrated into our CI. >>> >>> Right, I should have modified this chapter to be coherent with your latest tests. >>> Sorry for that. >>> >>>> >>>> I will do some testing and get back to you with results + review. >> I did some testing and here are the results: >> >> In the current form this series will fail when running CI because the Yocto containers >> are based on "From ubuntu:22.04" (there is no platform prefix), which means that the containers >> are built for the host architecture (in my case and in 99% of the cases of the local build it will >> be x86). In Gitlab we have 2 runners (arm64 and x86_64). This means that all the test jobs would need >> to specify x86_64 as a tag when keeping the current behavior. >> After I built all the containers on my x86 machine, I pushed them to registry and the pipeline was successful: >> https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgitlab.com%2Fxen-project%2Fpeople%2Fmorzel%2Fxen-orzelmichal%2F-%2Fpipelines%2F686853939&data=05%7C01%7Cmichal.orzel%40amd.com%7C097e7825c8834c412e8008dac16526bf%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C638034937437190896%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=MNo1wYRgxehJ0SJOeGxNR6U%2Fd68zERPgYfLbpHya%2B2Y%3D&reserved=0 >> >> Here is the diff on patch no. 3 to make the series work (using x86 tag and small improvement to include needs: []): >> ``` >> diff --git a/automation/gitlab-ci/test.yaml b/automation/gitlab-ci/test.yaml >> index 5c620fefce59..52cccec6f904 100644 >> --- a/automation/gitlab-ci/test.yaml >> +++ b/automation/gitlab-ci/test.yaml >> @@ -65,6 +65,9 @@ >> paths: >> - 'logs/*' >> when: always >> + needs: [] >> + tags: >> + - x86_64 >> >> # Test jobs >> build-each-commit-gcc: >> @@ -206,19 +209,13 @@ yocto-qemuarm64: >> extends: .yocto-test >> variables: >> YOCTO_BOARD: qemuarm64 >> - tags: >> - - arm64 >> >> yocto-qemuarm: >> extends: .yocto-test >> variables: >> YOCTO_BOARD: qemuarm >> - tags: >> - - arm32 >> >> yocto-qemux86-64: >> extends: .yocto-test >> variables: >> YOCTO_BOARD: qemux86-64 >> - tags: >> - - x86_64 >> ``` >> >> Now, the logical way would be to build x86 yocto container for x86, arm64 for arm64 and arm32 on arm64 or x86. >> I tried building the container qemuarm64 specifying target arm64 on x86. After 15h, only 70% of the Yocto build >> was completed and there was an error with glibc (the local build of the container for the host arch takes on my machine max 2h). >> This enormous amount of time is due to the qemu docker emulation that happens behind the scenes (I checked on 2 different machines). >> >> So we have 3 solutions: >> 1) Build and run these containers for/on x86_64: >> - local users can build the containers on local machines that are almost always x86 based, in short period of time, >> - "everyone" can build/push the containers once there is a new Yocto release >> - slightly slower CI build time >> 2) Build and run these containers for specific architectures: >> - almost no go for local users using x86 machine (unless using more than 16 threads (which I used) and willing to wait 2 days for the build) >> - faster CI build time (arm64 runner is faster than x86 one) >> - someone with arm64 based machine (not that common) would have to build and push the containers >> 3) Try to use CI to build and push the containers to registry >> - it could be possible but what about local users > > I might have a solution 4. > The containers we pushed to gitlab are publicly accessible ? Yes, they are: https://gitlab.com/xen-project/xen/container_registry/3539006 > Shouldn’t we push both versions there and then document how someone can retrieve and use those containers instead of building them ? > WRT retrieving, we already have a section describing it under automation/build/README.md I'm not sure about pushing both versions to registry, due to the following: - occupying twice the same size (instead of 30GB per release we would have 60GB, imagine the size after a few years), - necessity to build twice the same containers for each release (someone would have to have access to both x86 and arm64 machines to do the build), - necessity to introduce different naming for containers (to differentiate the machine) where in fact, CI would always use one subset. We should not worry about the local users, because once the containers are in registry (no matter the platform they were built for), downloading the containers will take less time than building them. We will not benefit anything from having containers for different platforms. The only problem is with building them on x86 for arm64 (other way around is not supported by qemu-user-static). But this is something that may be problematic only for person willing to push to registry, because as you said, the normal users can pull them instead of build them. > This way we will do the push but a normal user would just retrieve the containers and use them so building time would have less impact and we could build both flavours of the containers. > >> >> Regardless of what we chose, we need to keep in mind that the biggest advantage to the Yocto build/run is that >> it allows/should allow local users to perform basic testing for all the Xen supported architectures. This is because >> everything happens in one place with one command. > > Agree. > > Bertrand > >> >> ~Michal > ~Michal
On Mon, 7 Nov 2022, Michal Orzel wrote: > Hi Bertrand and Stefano, > > On 31/10/2022 16:00, Bertrand Marquis wrote: > > > > > > Hi Michal, > > > >> On 31 Oct 2022, at 14:39, Michal Orzel <michal.orzel@amd.com> wrote: > >> > >> Hi Bertrand, > >> > >> On 31/10/2022 15:00, Bertrand Marquis wrote: > >>> > >>> > >>> This patch series is a first attempt to check if we could use Yocto in > >>> gitlab ci to build and run xen on qemu for arm, arm64 and x86. > >>> > >>> The first patch is creating a container with all elements required to > >>> build Yocto, a checkout of the yocto layers required and an helper > >>> script to build and run xen on qemu with yocto. > >>> > >>> The second patch is creating containers with a first build of yocto done > >>> so that susbsequent build with those containers would only rebuild what > >>> was changed and take the rest from the cache. > >>> > >>> The third patch is adding a way to easily clean locally created > >>> containers. > >>> > >>> This is is mainly for discussion and sharing as there are still some > >>> issues/problem to solve: > >>> - building the qemu* containers can take several hours depending on the > >>> network bandwith and computing power of the machine where those are > >>> created > >> This is not really an issue as the build of the containers occurs on the local > >> machines before pushing them to registry. Also, building the containers > >> will only be required for new Yocto releases. > >> > >>> - produced containers containing the cache have a size between 8 and > >>> 12GB depending on the architecture. We might need to store the build > >>> cache somewhere else to reduce the size. If we choose to have one > >>> single image, the needed size is around 20GB and we need up to 40GB > >>> during the build, which is why I splitted them. > >>> - during the build and run, we use a bit more then 20GB of disk which is > >>> over the allowed size in gitlab > >> As we could see during v2 testing, we do not have any space restrictions > >> on the Xen GitLab and I think we already decided to have the Yocto > >> integrated into our CI. > > > > Right, I should have modified this chapter to be coherent with your latest tests. > > Sorry for that. > > > >> > >> I will do some testing and get back to you with results + review. > I did some testing and here are the results: > > In the current form this series will fail when running CI because the Yocto containers > are based on "From ubuntu:22.04" (there is no platform prefix), which means that the containers > are built for the host architecture (in my case and in 99% of the cases of the local build it will > be x86). In Gitlab we have 2 runners (arm64 and x86_64). This means that all the test jobs would need > to specify x86_64 as a tag when keeping the current behavior. > After I built all the containers on my x86 machine, I pushed them to registry and the pipeline was successful: > https://gitlab.com/xen-project/people/morzel/xen-orzelmichal/-/pipelines/686853939 When I tested the previous version of this series I built the containers natively on ARM64, so that is also an option. > Here is the diff on patch no. 3 to make the series work (using x86 tag and small improvement to include needs: []): > ``` > diff --git a/automation/gitlab-ci/test.yaml b/automation/gitlab-ci/test.yaml > index 5c620fefce59..52cccec6f904 100644 > --- a/automation/gitlab-ci/test.yaml > +++ b/automation/gitlab-ci/test.yaml > @@ -65,6 +65,9 @@ > paths: > - 'logs/*' > when: always > + needs: [] > + tags: > + - x86_64 > > # Test jobs > build-each-commit-gcc: > @@ -206,19 +209,13 @@ yocto-qemuarm64: > extends: .yocto-test > variables: > YOCTO_BOARD: qemuarm64 > - tags: > - - arm64 > > yocto-qemuarm: > extends: .yocto-test > variables: > YOCTO_BOARD: qemuarm > - tags: > - - arm32 > > yocto-qemux86-64: > extends: .yocto-test > variables: > YOCTO_BOARD: qemux86-64 > - tags: > - - x86_64 > ``` > > Now, the logical way would be to build x86 yocto container for x86, arm64 for arm64 and arm32 on arm64 or x86. > I tried building the container qemuarm64 specifying target arm64 on x86. After 15h, only 70% of the Yocto build > was completed and there was an error with glibc (the local build of the container for the host arch takes on my machine max 2h). > This enormous amount of time is due to the qemu docker emulation that happens behind the scenes (I checked on 2 different machines). > > So we have 3 solutions: > 1) Build and run these containers for/on x86_64: > - local users can build the containers on local machines that are almost always x86 based, in short period of time, > - "everyone" can build/push the containers once there is a new Yocto release > - slightly slower CI build time > 2) Build and run these containers for specific architectures: > - almost no go for local users using x86 machine (unless using more than 16 threads (which I used) and willing to wait 2 days for the build) > - faster CI build time (arm64 runner is faster than x86 one) > - someone with arm64 based machine (not that common) would have to build and push the containers > 3) Try to use CI to build and push the containers to registry > - it could be possible but what about local users From a gitlab-ci perspective, given the runners we currently have, we have to go with option 2). We don't have enough resources available on the x86 runner to run the Yocto jobs on x86. > Regardless of what we chose, we need to keep in mind that the biggest advantage to the Yocto build/run is that > it allows/should allow local users to perform basic testing for all the Xen supported architectures. This is because > everything happens in one place with one command. That's right, but it should be possible to allow the Yocto containers to also build and run correctly locally on x86, right? The arm/x86 tag in test.yaml doesn't matter when running the containers locally anyway.
Hi Stefano, On 10/11/2022 01:18, Stefano Stabellini wrote: > > > On Mon, 7 Nov 2022, Michal Orzel wrote: >> Hi Bertrand and Stefano, >> >> On 31/10/2022 16:00, Bertrand Marquis wrote: >>> >>> >>> Hi Michal, >>> >>>> On 31 Oct 2022, at 14:39, Michal Orzel <michal.orzel@amd.com> wrote: >>>> >>>> Hi Bertrand, >>>> >>>> On 31/10/2022 15:00, Bertrand Marquis wrote: >>>>> >>>>> >>>>> This patch series is a first attempt to check if we could use Yocto in >>>>> gitlab ci to build and run xen on qemu for arm, arm64 and x86. >>>>> >>>>> The first patch is creating a container with all elements required to >>>>> build Yocto, a checkout of the yocto layers required and an helper >>>>> script to build and run xen on qemu with yocto. >>>>> >>>>> The second patch is creating containers with a first build of yocto done >>>>> so that susbsequent build with those containers would only rebuild what >>>>> was changed and take the rest from the cache. >>>>> >>>>> The third patch is adding a way to easily clean locally created >>>>> containers. >>>>> >>>>> This is is mainly for discussion and sharing as there are still some >>>>> issues/problem to solve: >>>>> - building the qemu* containers can take several hours depending on the >>>>> network bandwith and computing power of the machine where those are >>>>> created >>>> This is not really an issue as the build of the containers occurs on the local >>>> machines before pushing them to registry. Also, building the containers >>>> will only be required for new Yocto releases. >>>> >>>>> - produced containers containing the cache have a size between 8 and >>>>> 12GB depending on the architecture. We might need to store the build >>>>> cache somewhere else to reduce the size. If we choose to have one >>>>> single image, the needed size is around 20GB and we need up to 40GB >>>>> during the build, which is why I splitted them. >>>>> - during the build and run, we use a bit more then 20GB of disk which is >>>>> over the allowed size in gitlab >>>> As we could see during v2 testing, we do not have any space restrictions >>>> on the Xen GitLab and I think we already decided to have the Yocto >>>> integrated into our CI. >>> >>> Right, I should have modified this chapter to be coherent with your latest tests. >>> Sorry for that. >>> >>>> >>>> I will do some testing and get back to you with results + review. >> I did some testing and here are the results: >> >> In the current form this series will fail when running CI because the Yocto containers >> are based on "From ubuntu:22.04" (there is no platform prefix), which means that the containers >> are built for the host architecture (in my case and in 99% of the cases of the local build it will >> be x86). In Gitlab we have 2 runners (arm64 and x86_64). This means that all the test jobs would need >> to specify x86_64 as a tag when keeping the current behavior. >> After I built all the containers on my x86 machine, I pushed them to registry and the pipeline was successful: >> https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgitlab.com%2Fxen-project%2Fpeople%2Fmorzel%2Fxen-orzelmichal%2F-%2Fpipelines%2F686853939&data=05%7C01%7Cmichal.orzel%40amd.com%7C2449f063e67341c3b95a08dac2b112a5%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C638036363027707274%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=EwTJrW2vuwQIugKc7mnzG9NNbsYLP6tw5UODzBMmPEE%3D&reserved=0 > > When I tested the previous version of this series I built the > containers natively on ARM64, so that is also an option. > > >> Here is the diff on patch no. 3 to make the series work (using x86 tag and small improvement to include needs: []): >> ``` >> diff --git a/automation/gitlab-ci/test.yaml b/automation/gitlab-ci/test.yaml >> index 5c620fefce59..52cccec6f904 100644 >> --- a/automation/gitlab-ci/test.yaml >> +++ b/automation/gitlab-ci/test.yaml >> @@ -65,6 +65,9 @@ >> paths: >> - 'logs/*' >> when: always >> + needs: [] >> + tags: >> + - x86_64 >> >> # Test jobs >> build-each-commit-gcc: >> @@ -206,19 +209,13 @@ yocto-qemuarm64: >> extends: .yocto-test >> variables: >> YOCTO_BOARD: qemuarm64 >> - tags: >> - - arm64 >> >> yocto-qemuarm: >> extends: .yocto-test >> variables: >> YOCTO_BOARD: qemuarm >> - tags: >> - - arm32 >> >> yocto-qemux86-64: >> extends: .yocto-test >> variables: >> YOCTO_BOARD: qemux86-64 >> - tags: >> - - x86_64 >> ``` >> >> Now, the logical way would be to build x86 yocto container for x86, arm64 for arm64 and arm32 on arm64 or x86. >> I tried building the container qemuarm64 specifying target arm64 on x86. After 15h, only 70% of the Yocto build >> was completed and there was an error with glibc (the local build of the container for the host arch takes on my machine max 2h). >> This enormous amount of time is due to the qemu docker emulation that happens behind the scenes (I checked on 2 different machines). >> >> So we have 3 solutions: >> 1) Build and run these containers for/on x86_64: >> - local users can build the containers on local machines that are almost always x86 based, in short period of time, >> - "everyone" can build/push the containers once there is a new Yocto release >> - slightly slower CI build time >> 2) Build and run these containers for specific architectures: >> - almost no go for local users using x86 machine (unless using more than 16 threads (which I used) and willing to wait 2 days for the build) >> - faster CI build time (arm64 runner is faster than x86 one) >> - someone with arm64 based machine (not that common) would have to build and push the containers >> 3) Try to use CI to build and push the containers to registry >> - it could be possible but what about local users > > From a gitlab-ci perspective, given the runners we currently have, we > have to go with option 2). We don't have enough resources available on > the x86 runner to run the Yocto jobs on x86. > That is what I reckon too. Running the Yocto build/test on CI using x86 runner will always be slower. So, if we go with this solution, then the following is needed: 1. Modify test jobs so that yocto-qemu{arm64/arm} uses arm64 tag to be taken by arm64 runner and use tag x86_64 for yocto-qemux86-64. 2. Come up with a solution to build the yocto containers automatically for the above platforms + possibility to specify the platform for local users. Right now, these containers are being always build for the host machine platform, so without doing tricks like adding --platform or prefix to image name, one cannot build the Yocto containers that would be ready to be pushed to registry. We need to have a clean solution without requiring user to do tricks. The only drawback of this solution is that the person building the yocto-qemu{arm64/arm} container and willing to push it to registry, needs to have access to arm64 machine. > >> Regardless of what we chose, we need to keep in mind that the biggest advantage to the Yocto build/run is that >> it allows/should allow local users to perform basic testing for all the Xen supported architectures. This is because >> everything happens in one place with one command. > > That's right, but it should be possible to allow the Yocto containers to > also build and run correctly locally on x86, right? The arm/x86 tag in > test.yaml doesn't matter when running the containers locally anyway. ~Michal
Hi Michal, > On 10 Nov 2022, at 07:34, Michal Orzel <michal.orzel@amd.com> wrote: > > Hi Stefano, > > On 10/11/2022 01:18, Stefano Stabellini wrote: >> >> >> On Mon, 7 Nov 2022, Michal Orzel wrote: >>> Hi Bertrand and Stefano, >>> >>> On 31/10/2022 16:00, Bertrand Marquis wrote: >>>> >>>> >>>> Hi Michal, >>>> >>>>> On 31 Oct 2022, at 14:39, Michal Orzel <michal.orzel@amd.com> wrote: >>>>> >>>>> Hi Bertrand, >>>>> >>>>> On 31/10/2022 15:00, Bertrand Marquis wrote: >>>>>> >>>>>> >>>>>> This patch series is a first attempt to check if we could use Yocto in >>>>>> gitlab ci to build and run xen on qemu for arm, arm64 and x86. >>>>>> >>>>>> The first patch is creating a container with all elements required to >>>>>> build Yocto, a checkout of the yocto layers required and an helper >>>>>> script to build and run xen on qemu with yocto. >>>>>> >>>>>> The second patch is creating containers with a first build of yocto done >>>>>> so that susbsequent build with those containers would only rebuild what >>>>>> was changed and take the rest from the cache. >>>>>> >>>>>> The third patch is adding a way to easily clean locally created >>>>>> containers. >>>>>> >>>>>> This is is mainly for discussion and sharing as there are still some >>>>>> issues/problem to solve: >>>>>> - building the qemu* containers can take several hours depending on the >>>>>> network bandwith and computing power of the machine where those are >>>>>> created >>>>> This is not really an issue as the build of the containers occurs on the local >>>>> machines before pushing them to registry. Also, building the containers >>>>> will only be required for new Yocto releases. >>>>> >>>>>> - produced containers containing the cache have a size between 8 and >>>>>> 12GB depending on the architecture. We might need to store the build >>>>>> cache somewhere else to reduce the size. If we choose to have one >>>>>> single image, the needed size is around 20GB and we need up to 40GB >>>>>> during the build, which is why I splitted them. >>>>>> - during the build and run, we use a bit more then 20GB of disk which is >>>>>> over the allowed size in gitlab >>>>> As we could see during v2 testing, we do not have any space restrictions >>>>> on the Xen GitLab and I think we already decided to have the Yocto >>>>> integrated into our CI. >>>> >>>> Right, I should have modified this chapter to be coherent with your latest tests. >>>> Sorry for that. >>>> >>>>> >>>>> I will do some testing and get back to you with results + review. >>> I did some testing and here are the results: >>> >>> In the current form this series will fail when running CI because the Yocto containers >>> are based on "From ubuntu:22.04" (there is no platform prefix), which means that the containers >>> are built for the host architecture (in my case and in 99% of the cases of the local build it will >>> be x86). In Gitlab we have 2 runners (arm64 and x86_64). This means that all the test jobs would need >>> to specify x86_64 as a tag when keeping the current behavior. >>> After I built all the containers on my x86 machine, I pushed them to registry and the pipeline was successful: >>> https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgitlab.com%2Fxen-project%2Fpeople%2Fmorzel%2Fxen-orzelmichal%2F-%2Fpipelines%2F686853939&data=05%7C01%7Cmichal.orzel%40amd.com%7C2449f063e67341c3b95a08dac2b112a5%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C638036363027707274%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=EwTJrW2vuwQIugKc7mnzG9NNbsYLP6tw5UODzBMmPEE%3D&reserved=0 >> >> When I tested the previous version of this series I built the >> containers natively on ARM64, so that is also an option. >> >> >>> Here is the diff on patch no. 3 to make the series work (using x86 tag and small improvement to include needs: []): >>> ``` >>> diff --git a/automation/gitlab-ci/test.yaml b/automation/gitlab-ci/test.yaml >>> index 5c620fefce59..52cccec6f904 100644 >>> --- a/automation/gitlab-ci/test.yaml >>> +++ b/automation/gitlab-ci/test.yaml >>> @@ -65,6 +65,9 @@ >>> paths: >>> - 'logs/*' >>> when: always >>> + needs: [] >>> + tags: >>> + - x86_64 >>> >>> # Test jobs >>> build-each-commit-gcc: >>> @@ -206,19 +209,13 @@ yocto-qemuarm64: >>> extends: .yocto-test >>> variables: >>> YOCTO_BOARD: qemuarm64 >>> - tags: >>> - - arm64 >>> >>> yocto-qemuarm: >>> extends: .yocto-test >>> variables: >>> YOCTO_BOARD: qemuarm >>> - tags: >>> - - arm32 >>> >>> yocto-qemux86-64: >>> extends: .yocto-test >>> variables: >>> YOCTO_BOARD: qemux86-64 >>> - tags: >>> - - x86_64 >>> ``` >>> >>> Now, the logical way would be to build x86 yocto container for x86, arm64 for arm64 and arm32 on arm64 or x86. >>> I tried building the container qemuarm64 specifying target arm64 on x86. After 15h, only 70% of the Yocto build >>> was completed and there was an error with glibc (the local build of the container for the host arch takes on my machine max 2h). >>> This enormous amount of time is due to the qemu docker emulation that happens behind the scenes (I checked on 2 different machines). >>> >>> So we have 3 solutions: >>> 1) Build and run these containers for/on x86_64: >>> - local users can build the containers on local machines that are almost always x86 based, in short period of time, >>> - "everyone" can build/push the containers once there is a new Yocto release >>> - slightly slower CI build time >>> 2) Build and run these containers for specific architectures: >>> - almost no go for local users using x86 machine (unless using more than 16 threads (which I used) and willing to wait 2 days for the build) >>> - faster CI build time (arm64 runner is faster than x86 one) >>> - someone with arm64 based machine (not that common) would have to build and push the containers >>> 3) Try to use CI to build and push the containers to registry >>> - it could be possible but what about local users >> >> From a gitlab-ci perspective, given the runners we currently have, we >> have to go with option 2). We don't have enough resources available on >> the x86 runner to run the Yocto jobs on x86. >> > That is what I reckon too. Running the Yocto build/test on CI using x86 runner will always be slower. > So, if we go with this solution, then the following is needed: > 1. Modify test jobs so that yocto-qemu{arm64/arm} uses arm64 tag to be taken by arm64 runner and use tag x86_64 for yocto-qemux86-64. > 2. Come up with a solution to build the yocto containers automatically for the above platforms + possibility to specify the platform for local users. > Right now, these containers are being always build for the host machine platform, so without doing tricks like adding --platform or prefix to image name, > one cannot build the Yocto containers that would be ready to be pushed to registry. We need to have a clean solution without requiring user to do tricks. > > The only drawback of this solution is that the person building the yocto-qemu{arm64/arm} container and willing to push it to registry, > needs to have access to arm64 machine. I think we need to find a solution working for both possibilities. And we also need a solution so that one can have both kind of images so the host machine should be encoded in the container name somehow. > >> >>> Regardless of what we chose, we need to keep in mind that the biggest advantage to the Yocto build/run is that >>> it allows/should allow local users to perform basic testing for all the Xen supported architectures. This is because >>> everything happens in one place with one command. >> >> That's right, but it should be possible to allow the Yocto containers to >> also build and run correctly locally on x86, right? The arm/x86 tag in >> test.yaml doesn't matter when running the containers locally anyway. All in all, test.yaml only matter for gitlab. Maybe we could have it supporting both cases but only use one ? Cheers Bertrand > > ~Michal
Hello, On 10/11/2022 11:08, Bertrand Marquis wrote: > > > Hi Michal, > >> On 10 Nov 2022, at 07:34, Michal Orzel <michal.orzel@amd.com> wrote: >> >> Hi Stefano, >> >> On 10/11/2022 01:18, Stefano Stabellini wrote: >>> >>> >>> On Mon, 7 Nov 2022, Michal Orzel wrote: >>>> Hi Bertrand and Stefano, >>>> >>>> On 31/10/2022 16:00, Bertrand Marquis wrote: >>>>> >>>>> >>>>> Hi Michal, >>>>> >>>>>> On 31 Oct 2022, at 14:39, Michal Orzel <michal.orzel@amd.com> wrote: >>>>>> >>>>>> Hi Bertrand, >>>>>> >>>>>> On 31/10/2022 15:00, Bertrand Marquis wrote: >>>>>>> >>>>>>> >>>>>>> This patch series is a first attempt to check if we could use Yocto in >>>>>>> gitlab ci to build and run xen on qemu for arm, arm64 and x86. >>>>>>> >>>>>>> The first patch is creating a container with all elements required to >>>>>>> build Yocto, a checkout of the yocto layers required and an helper >>>>>>> script to build and run xen on qemu with yocto. >>>>>>> >>>>>>> The second patch is creating containers with a first build of yocto done >>>>>>> so that susbsequent build with those containers would only rebuild what >>>>>>> was changed and take the rest from the cache. >>>>>>> >>>>>>> The third patch is adding a way to easily clean locally created >>>>>>> containers. >>>>>>> >>>>>>> This is is mainly for discussion and sharing as there are still some >>>>>>> issues/problem to solve: >>>>>>> - building the qemu* containers can take several hours depending on the >>>>>>> network bandwith and computing power of the machine where those are >>>>>>> created >>>>>> This is not really an issue as the build of the containers occurs on the local >>>>>> machines before pushing them to registry. Also, building the containers >>>>>> will only be required for new Yocto releases. >>>>>> >>>>>>> - produced containers containing the cache have a size between 8 and >>>>>>> 12GB depending on the architecture. We might need to store the build >>>>>>> cache somewhere else to reduce the size. If we choose to have one >>>>>>> single image, the needed size is around 20GB and we need up to 40GB >>>>>>> during the build, which is why I splitted them. >>>>>>> - during the build and run, we use a bit more then 20GB of disk which is >>>>>>> over the allowed size in gitlab >>>>>> As we could see during v2 testing, we do not have any space restrictions >>>>>> on the Xen GitLab and I think we already decided to have the Yocto >>>>>> integrated into our CI. >>>>> >>>>> Right, I should have modified this chapter to be coherent with your latest tests. >>>>> Sorry for that. >>>>> >>>>>> >>>>>> I will do some testing and get back to you with results + review. >>>> I did some testing and here are the results: >>>> >>>> In the current form this series will fail when running CI because the Yocto containers >>>> are based on "From ubuntu:22.04" (there is no platform prefix), which means that the containers >>>> are built for the host architecture (in my case and in 99% of the cases of the local build it will >>>> be x86). In Gitlab we have 2 runners (arm64 and x86_64). This means that all the test jobs would need >>>> to specify x86_64 as a tag when keeping the current behavior. >>>> After I built all the containers on my x86 machine, I pushed them to registry and the pipeline was successful: >>>> https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgitlab.com%2Fxen-project%2Fpeople%2Fmorzel%2Fxen-orzelmichal%2F-%2Fpipelines%2F686853939&data=05%7C01%7Cmichal.orzel%40amd.com%7Ccc0a420856c64224e78208dac3037b95%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C638036716985754450%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=ARpspaYQlf7gujC11cS4dnR275cCJ6tMRC2J0FyhFIM%3D&reserved=0 >>> >>> When I tested the previous version of this series I built the >>> containers natively on ARM64, so that is also an option. >>> >>> >>>> Here is the diff on patch no. 3 to make the series work (using x86 tag and small improvement to include needs: []): >>>> ``` >>>> diff --git a/automation/gitlab-ci/test.yaml b/automation/gitlab-ci/test.yaml >>>> index 5c620fefce59..52cccec6f904 100644 >>>> --- a/automation/gitlab-ci/test.yaml >>>> +++ b/automation/gitlab-ci/test.yaml >>>> @@ -65,6 +65,9 @@ >>>> paths: >>>> - 'logs/*' >>>> when: always >>>> + needs: [] >>>> + tags: >>>> + - x86_64 >>>> >>>> # Test jobs >>>> build-each-commit-gcc: >>>> @@ -206,19 +209,13 @@ yocto-qemuarm64: >>>> extends: .yocto-test >>>> variables: >>>> YOCTO_BOARD: qemuarm64 >>>> - tags: >>>> - - arm64 >>>> >>>> yocto-qemuarm: >>>> extends: .yocto-test >>>> variables: >>>> YOCTO_BOARD: qemuarm >>>> - tags: >>>> - - arm32 >>>> >>>> yocto-qemux86-64: >>>> extends: .yocto-test >>>> variables: >>>> YOCTO_BOARD: qemux86-64 >>>> - tags: >>>> - - x86_64 >>>> ``` >>>> >>>> Now, the logical way would be to build x86 yocto container for x86, arm64 for arm64 and arm32 on arm64 or x86. >>>> I tried building the container qemuarm64 specifying target arm64 on x86. After 15h, only 70% of the Yocto build >>>> was completed and there was an error with glibc (the local build of the container for the host arch takes on my machine max 2h). >>>> This enormous amount of time is due to the qemu docker emulation that happens behind the scenes (I checked on 2 different machines). >>>> >>>> So we have 3 solutions: >>>> 1) Build and run these containers for/on x86_64: >>>> - local users can build the containers on local machines that are almost always x86 based, in short period of time, >>>> - "everyone" can build/push the containers once there is a new Yocto release >>>> - slightly slower CI build time >>>> 2) Build and run these containers for specific architectures: >>>> - almost no go for local users using x86 machine (unless using more than 16 threads (which I used) and willing to wait 2 days for the build) >>>> - faster CI build time (arm64 runner is faster than x86 one) >>>> - someone with arm64 based machine (not that common) would have to build and push the containers >>>> 3) Try to use CI to build and push the containers to registry >>>> - it could be possible but what about local users >>> >>> From a gitlab-ci perspective, given the runners we currently have, we >>> have to go with option 2). We don't have enough resources available on >>> the x86 runner to run the Yocto jobs on x86. >>> >> That is what I reckon too. Running the Yocto build/test on CI using x86 runner will always be slower. >> So, if we go with this solution, then the following is needed: >> 1. Modify test jobs so that yocto-qemu{arm64/arm} uses arm64 tag to be taken by arm64 runner and use tag x86_64 for yocto-qemux86-64. >> 2. Come up with a solution to build the yocto containers automatically for the above platforms + possibility to specify the platform for local users. >> Right now, these containers are being always build for the host machine platform, so without doing tricks like adding --platform or prefix to image name, >> one cannot build the Yocto containers that would be ready to be pushed to registry. We need to have a clean solution without requiring user to do tricks. >> >> The only drawback of this solution is that the person building the yocto-qemu{arm64/arm} container and willing to push it to registry, >> needs to have access to arm64 machine. > > I think we need to find a solution working for both possibilities. > And we also need a solution so that one can have both kind of images so the host machine should be encoded in the container name somehow. > >> >>> >>>> Regardless of what we chose, we need to keep in mind that the biggest advantage to the Yocto build/run is that >>>> it allows/should allow local users to perform basic testing for all the Xen supported architectures. This is because >>>> everything happens in one place with one command. >>> >>> That's right, but it should be possible to allow the Yocto containers to >>> also build and run correctly locally on x86, right? The arm/x86 tag in >>> test.yaml doesn't matter when running the containers locally anyway. > > All in all, test.yaml only matter for gitlab. > Maybe we could have it supporting both cases but only use one ? > > Cheers > Bertrand > >> >> ~Michal > I think we are trying to over engineer this problem. We want to start using it asap and this is what we should care about. If we start going into details, like adding more and more code (e.g take into consideration the fact that qemu emulation works only on x86 and what about users having arm64 machine), we will start creating a mess. CURRENT BEHAVIOR: All the containers are build for the host machine architecture. For me, it can stay as a default behavior so that all the local users can make use of it. WHAT WE WANT: The default behavior like above and the possibility to specify the platform the containers will be build for. This is so that we can build the containers for arm on x86 and use it for registry to be used by CI. RESULT: Default behavior is used by local users allowing to perform relatively quick build and test. Possibility to specify the platform is used by the person building and pushing the containers for CI runners. We do not need any duplication in containers' naming or in test.yaml test jobs. Test jobs specify tags to link the job with a proper runner and that is why the registry should only have containers built for proper runners' architectures. ~Michal
On Mon, Nov 07, 2022 at 08:50:09AM +0100, Michal Orzel wrote: > 3) Try to use CI to build and push the containers to registry > - it could be possible but what about local users FYI, it's already possible to build and push container from the CI, at least on X86, I've made it work: https://lore.kernel.org/all/20220301121133.19271-3-anthony.perard@citrix.com/ This works only when pushing to the "staging" branch on the main "xen" repo as I've added secret variable there. (Also the branch needs to be "protected" if I remember correctly.) I don't know how to use the Arm runner to do container builds like that. (The change done to one x86 runner: https://gitlab.com/xen-project/xen-gitlab-ci/-/merge_requests/15) Cheers,
Hi Anthony, > On 10 Nov 2022, at 13:20, Anthony PERARD <anthony.perard@citrix.com> wrote: > > On Mon, Nov 07, 2022 at 08:50:09AM +0100, Michal Orzel wrote: >> 3) Try to use CI to build and push the containers to registry >> - it could be possible but what about local users > > FYI, it's already possible to build and push container from the CI, at > least on X86, I've made it work: > https://lore.kernel.org/all/20220301121133.19271-3-anthony.perard@citrix.com/ > This works only when pushing to the "staging" branch on the main "xen" > repo as I've added secret variable there. (Also the branch needs to be > "protected" if I remember correctly.) Very nice :-) Would definitely be a good solution to have a solution like that for yocto. > > I don't know how to use the Arm runner to do container builds like that. > > (The change done to one x86 runner: > https://gitlab.com/xen-project/xen-gitlab-ci/-/merge_requests/15) Something to look at definitely. Cheers Bertrand > > Cheers, > > -- > Anthony PERARD
On Thu, 10 Nov 2022, Michal Orzel wrote: > Hi Stefano, > > On 10/11/2022 01:18, Stefano Stabellini wrote: > > > > > > On Mon, 7 Nov 2022, Michal Orzel wrote: > >> Hi Bertrand and Stefano, > >> > >> On 31/10/2022 16:00, Bertrand Marquis wrote: > >>> > >>> > >>> Hi Michal, > >>> > >>>> On 31 Oct 2022, at 14:39, Michal Orzel <michal.orzel@amd.com> wrote: > >>>> > >>>> Hi Bertrand, > >>>> > >>>> On 31/10/2022 15:00, Bertrand Marquis wrote: > >>>>> > >>>>> > >>>>> This patch series is a first attempt to check if we could use Yocto in > >>>>> gitlab ci to build and run xen on qemu for arm, arm64 and x86. > >>>>> > >>>>> The first patch is creating a container with all elements required to > >>>>> build Yocto, a checkout of the yocto layers required and an helper > >>>>> script to build and run xen on qemu with yocto. > >>>>> > >>>>> The second patch is creating containers with a first build of yocto done > >>>>> so that susbsequent build with those containers would only rebuild what > >>>>> was changed and take the rest from the cache. > >>>>> > >>>>> The third patch is adding a way to easily clean locally created > >>>>> containers. > >>>>> > >>>>> This is is mainly for discussion and sharing as there are still some > >>>>> issues/problem to solve: > >>>>> - building the qemu* containers can take several hours depending on the > >>>>> network bandwith and computing power of the machine where those are > >>>>> created > >>>> This is not really an issue as the build of the containers occurs on the local > >>>> machines before pushing them to registry. Also, building the containers > >>>> will only be required for new Yocto releases. > >>>> > >>>>> - produced containers containing the cache have a size between 8 and > >>>>> 12GB depending on the architecture. We might need to store the build > >>>>> cache somewhere else to reduce the size. If we choose to have one > >>>>> single image, the needed size is around 20GB and we need up to 40GB > >>>>> during the build, which is why I splitted them. > >>>>> - during the build and run, we use a bit more then 20GB of disk which is > >>>>> over the allowed size in gitlab > >>>> As we could see during v2 testing, we do not have any space restrictions > >>>> on the Xen GitLab and I think we already decided to have the Yocto > >>>> integrated into our CI. > >>> > >>> Right, I should have modified this chapter to be coherent with your latest tests. > >>> Sorry for that. > >>> > >>>> > >>>> I will do some testing and get back to you with results + review. > >> I did some testing and here are the results: > >> > >> In the current form this series will fail when running CI because the Yocto containers > >> are based on "From ubuntu:22.04" (there is no platform prefix), which means that the containers > >> are built for the host architecture (in my case and in 99% of the cases of the local build it will > >> be x86). In Gitlab we have 2 runners (arm64 and x86_64). This means that all the test jobs would need > >> to specify x86_64 as a tag when keeping the current behavior. > >> After I built all the containers on my x86 machine, I pushed them to registry and the pipeline was successful: > >> https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgitlab.com%2Fxen-project%2Fpeople%2Fmorzel%2Fxen-orzelmichal%2F-%2Fpipelines%2F686853939&data=05%7C01%7Cmichal.orzel%40amd.com%7C2449f063e67341c3b95a08dac2b112a5%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C638036363027707274%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=EwTJrW2vuwQIugKc7mnzG9NNbsYLP6tw5UODzBMmPEE%3D&reserved=0 > > > > When I tested the previous version of this series I built the > > containers natively on ARM64, so that is also an option. > > > > > >> Here is the diff on patch no. 3 to make the series work (using x86 tag and small improvement to include needs: []): > >> ``` > >> diff --git a/automation/gitlab-ci/test.yaml b/automation/gitlab-ci/test.yaml > >> index 5c620fefce59..52cccec6f904 100644 > >> --- a/automation/gitlab-ci/test.yaml > >> +++ b/automation/gitlab-ci/test.yaml > >> @@ -65,6 +65,9 @@ > >> paths: > >> - 'logs/*' > >> when: always > >> + needs: [] > >> + tags: > >> + - x86_64 > >> > >> # Test jobs > >> build-each-commit-gcc: > >> @@ -206,19 +209,13 @@ yocto-qemuarm64: > >> extends: .yocto-test > >> variables: > >> YOCTO_BOARD: qemuarm64 > >> - tags: > >> - - arm64 > >> > >> yocto-qemuarm: > >> extends: .yocto-test > >> variables: > >> YOCTO_BOARD: qemuarm > >> - tags: > >> - - arm32 > >> > >> yocto-qemux86-64: > >> extends: .yocto-test > >> variables: > >> YOCTO_BOARD: qemux86-64 > >> - tags: > >> - - x86_64 > >> ``` > >> > >> Now, the logical way would be to build x86 yocto container for x86, arm64 for arm64 and arm32 on arm64 or x86. > >> I tried building the container qemuarm64 specifying target arm64 on x86. After 15h, only 70% of the Yocto build > >> was completed and there was an error with glibc (the local build of the container for the host arch takes on my machine max 2h). > >> This enormous amount of time is due to the qemu docker emulation that happens behind the scenes (I checked on 2 different machines). > >> > >> So we have 3 solutions: > >> 1) Build and run these containers for/on x86_64: > >> - local users can build the containers on local machines that are almost always x86 based, in short period of time, > >> - "everyone" can build/push the containers once there is a new Yocto release > >> - slightly slower CI build time > >> 2) Build and run these containers for specific architectures: > >> - almost no go for local users using x86 machine (unless using more than 16 threads (which I used) and willing to wait 2 days for the build) > >> - faster CI build time (arm64 runner is faster than x86 one) > >> - someone with arm64 based machine (not that common) would have to build and push the containers > >> 3) Try to use CI to build and push the containers to registry > >> - it could be possible but what about local users > > > > From a gitlab-ci perspective, given the runners we currently have, we > > have to go with option 2). We don't have enough resources available on > > the x86 runner to run the Yocto jobs on x86. > > > That is what I reckon too. Running the Yocto build/test on CI using x86 runner will always be slower. > So, if we go with this solution, then the following is needed: > 1. Modify test jobs so that yocto-qemu{arm64/arm} uses arm64 tag to be taken by arm64 runner and use tag x86_64 for yocto-qemux86-64. > 2. Come up with a solution to build the yocto containers automatically for the above platforms + possibility to specify the platform for local users. > Right now, these containers are being always build for the host machine platform, so without doing tricks like adding --platform or prefix to image name, > one cannot build the Yocto containers that would be ready to be pushed to registry. We need to have a clean solution without requiring user to do tricks. > > The only drawback of this solution is that the person building the yocto-qemu{arm64/arm} container and willing to push it to registry, > needs to have access to arm64 machine. I am fine with this drawback for now. Due to resource constraints, we might want to avoid adding yocto-qemux86-64 (if yocto-qemux86-64 has to run on x86) for now, I worry it might choke the x86 ci-loop. Or we could add it but keep it disabled. We'll enable it when we get better x86 runners.
Hi Stefano, > On 10 Nov 2022, at 20:25, Stefano Stabellini <sstabellini@kernel.org> wrote: > > On Thu, 10 Nov 2022, Michal Orzel wrote: >> Hi Stefano, >> >> On 10/11/2022 01:18, Stefano Stabellini wrote: >>> >>> >>> On Mon, 7 Nov 2022, Michal Orzel wrote: >>>> Hi Bertrand and Stefano, >>>> >>>> On 31/10/2022 16:00, Bertrand Marquis wrote: >>>>> >>>>> >>>>> Hi Michal, >>>>> >>>>>> On 31 Oct 2022, at 14:39, Michal Orzel <michal.orzel@amd.com> wrote: >>>>>> >>>>>> Hi Bertrand, >>>>>> >>>>>> On 31/10/2022 15:00, Bertrand Marquis wrote: >>>>>>> >>>>>>> >>>>>>> This patch series is a first attempt to check if we could use Yocto in >>>>>>> gitlab ci to build and run xen on qemu for arm, arm64 and x86. >>>>>>> >>>>>>> The first patch is creating a container with all elements required to >>>>>>> build Yocto, a checkout of the yocto layers required and an helper >>>>>>> script to build and run xen on qemu with yocto. >>>>>>> >>>>>>> The second patch is creating containers with a first build of yocto done >>>>>>> so that susbsequent build with those containers would only rebuild what >>>>>>> was changed and take the rest from the cache. >>>>>>> >>>>>>> The third patch is adding a way to easily clean locally created >>>>>>> containers. >>>>>>> >>>>>>> This is is mainly for discussion and sharing as there are still some >>>>>>> issues/problem to solve: >>>>>>> - building the qemu* containers can take several hours depending on the >>>>>>> network bandwith and computing power of the machine where those are >>>>>>> created >>>>>> This is not really an issue as the build of the containers occurs on the local >>>>>> machines before pushing them to registry. Also, building the containers >>>>>> will only be required for new Yocto releases. >>>>>> >>>>>>> - produced containers containing the cache have a size between 8 and >>>>>>> 12GB depending on the architecture. We might need to store the build >>>>>>> cache somewhere else to reduce the size. If we choose to have one >>>>>>> single image, the needed size is around 20GB and we need up to 40GB >>>>>>> during the build, which is why I splitted them. >>>>>>> - during the build and run, we use a bit more then 20GB of disk which is >>>>>>> over the allowed size in gitlab >>>>>> As we could see during v2 testing, we do not have any space restrictions >>>>>> on the Xen GitLab and I think we already decided to have the Yocto >>>>>> integrated into our CI. >>>>> >>>>> Right, I should have modified this chapter to be coherent with your latest tests. >>>>> Sorry for that. >>>>> >>>>>> >>>>>> I will do some testing and get back to you with results + review. >>>> I did some testing and here are the results: >>>> >>>> In the current form this series will fail when running CI because the Yocto containers >>>> are based on "From ubuntu:22.04" (there is no platform prefix), which means that the containers >>>> are built for the host architecture (in my case and in 99% of the cases of the local build it will >>>> be x86). In Gitlab we have 2 runners (arm64 and x86_64). This means that all the test jobs would need >>>> to specify x86_64 as a tag when keeping the current behavior. >>>> After I built all the containers on my x86 machine, I pushed them to registry and the pipeline was successful: >>>> https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgitlab.com%2Fxen-project%2Fpeople%2Fmorzel%2Fxen-orzelmichal%2F-%2Fpipelines%2F686853939&data=05%7C01%7Cmichal.orzel%40amd.com%7C2449f063e67341c3b95a08dac2b112a5%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C638036363027707274%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=EwTJrW2vuwQIugKc7mnzG9NNbsYLP6tw5UODzBMmPEE%3D&reserved=0 >>> >>> When I tested the previous version of this series I built the >>> containers natively on ARM64, so that is also an option. >>> >>> >>>> Here is the diff on patch no. 3 to make the series work (using x86 tag and small improvement to include needs: []): >>>> ``` >>>> diff --git a/automation/gitlab-ci/test.yaml b/automation/gitlab-ci/test.yaml >>>> index 5c620fefce59..52cccec6f904 100644 >>>> --- a/automation/gitlab-ci/test.yaml >>>> +++ b/automation/gitlab-ci/test.yaml >>>> @@ -65,6 +65,9 @@ >>>> paths: >>>> - 'logs/*' >>>> when: always >>>> + needs: [] >>>> + tags: >>>> + - x86_64 >>>> >>>> # Test jobs >>>> build-each-commit-gcc: >>>> @@ -206,19 +209,13 @@ yocto-qemuarm64: >>>> extends: .yocto-test >>>> variables: >>>> YOCTO_BOARD: qemuarm64 >>>> - tags: >>>> - - arm64 >>>> >>>> yocto-qemuarm: >>>> extends: .yocto-test >>>> variables: >>>> YOCTO_BOARD: qemuarm >>>> - tags: >>>> - - arm32 >>>> >>>> yocto-qemux86-64: >>>> extends: .yocto-test >>>> variables: >>>> YOCTO_BOARD: qemux86-64 >>>> - tags: >>>> - - x86_64 >>>> ``` >>>> >>>> Now, the logical way would be to build x86 yocto container for x86, arm64 for arm64 and arm32 on arm64 or x86. >>>> I tried building the container qemuarm64 specifying target arm64 on x86. After 15h, only 70% of the Yocto build >>>> was completed and there was an error with glibc (the local build of the container for the host arch takes on my machine max 2h). >>>> This enormous amount of time is due to the qemu docker emulation that happens behind the scenes (I checked on 2 different machines). >>>> >>>> So we have 3 solutions: >>>> 1) Build and run these containers for/on x86_64: >>>> - local users can build the containers on local machines that are almost always x86 based, in short period of time, >>>> - "everyone" can build/push the containers once there is a new Yocto release >>>> - slightly slower CI build time >>>> 2) Build and run these containers for specific architectures: >>>> - almost no go for local users using x86 machine (unless using more than 16 threads (which I used) and willing to wait 2 days for the build) >>>> - faster CI build time (arm64 runner is faster than x86 one) >>>> - someone with arm64 based machine (not that common) would have to build and push the containers >>>> 3) Try to use CI to build and push the containers to registry >>>> - it could be possible but what about local users >>> >>> From a gitlab-ci perspective, given the runners we currently have, we >>> have to go with option 2). We don't have enough resources available on >>> the x86 runner to run the Yocto jobs on x86. >>> >> That is what I reckon too. Running the Yocto build/test on CI using x86 runner will always be slower. >> So, if we go with this solution, then the following is needed: >> 1. Modify test jobs so that yocto-qemu{arm64/arm} uses arm64 tag to be taken by arm64 runner and use tag x86_64 for yocto-qemux86-64. >> 2. Come up with a solution to build the yocto containers automatically for the above platforms + possibility to specify the platform for local users. >> Right now, these containers are being always build for the host machine platform, so without doing tricks like adding --platform or prefix to image name, >> one cannot build the Yocto containers that would be ready to be pushed to registry. We need to have a clean solution without requiring user to do tricks. >> >> The only drawback of this solution is that the person building the yocto-qemu{arm64/arm} container and willing to push it to registry, >> needs to have access to arm64 machine. > > I am fine with this drawback for now. > > Due to resource constraints, we might want to avoid adding > yocto-qemux86-64 (if yocto-qemux86-64 has to run on x86) for now, I > worry it might choke the x86 ci-loop. Or we could add it but keep it > disabled. We'll enable it when we get better x86 runners. Yocto-qemux86 actually runs quite well on an arm64 platform :-) Cheers Bertrand