I discovered that the master job is a red herring. The "error" is just a friendly notice that the first pipeline doesn't have the desired job. It successfully finds the job in the next pipeline it checks. Then, the runner ran out of disk space and the job died. Apparently I do not yet catch that particular spurious failure.
IIRC, the last time the same problem happened with a build-$release job, the solution was to twiddle the job name, wait for the problem to happen again, then twiddle it back. A release cycle ritual.
I think Matt's suggestion was to just check two names, x86_64-linux-fedora33-release and x86_64-linux-fedora33.
My preferred solution would be to make the same job have the same name regardless of which pipeline it's in, so maybe I'll open a ticket.
GitLab keeps the artifacts for the last pipeline on a ref indefinitely.
The last pipeline on release-9.6 was a Release pipeline, which doesn't have the right job.
The second-to-last pipeline is a Validate pipeline, which does have the right job.
Being the second-to-last pipeline, the artifacts had expired after some number of weeks.
I see Ben has answered at the same time as me, but I'll still post this since I expand on why it matters that the most recent pipeline isn't a Validate pipeline.
The ghc-9.6 failures are due to the fact that the most recent pipeline is a release pipeline, which does not have the needed validate artifact. I have started a new pipeline, https://gitlab.haskell.org/ghc/ghc/-/pipelines/65578, which should resolve this.
@mpickering suggests that the we may want to introduce a weekly validate job on the each of the stable branches to ensure that this doesn't happen again in the future (or at least place an upper bound on the amount of time for which the state will persist).
I guess this message slipped through the cracks. I am curious to know if this has been completely fixed, or if the 9.8 branch will somehow cause this problem to resurface.
Hard to say. I would be fine with closing this issue under the optimistic assumption that it has been fixed, allowing for the possibility of reopening it should the issue resurface.