-
Notifications
You must be signed in to change notification settings - Fork 1.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
flaky cross build #974
Comments
This is likely because we're launching many containers against this volume in parallel, which is mounted over a directory structure already containing sub-directories (and thus unioned). Without any mounts:
Locally this is typically not an issue because the volume has already been used. It might happen on clean builds, but it's going to be racy. The straightforward answer is to ensure that the directory / volume is fully setup before we start launching many parallel containers. |
#975 should fix this. |
I did some quick checks when I found this error. Under normal circumstances, there is no problem with creating volume repeatedly. I quickly searched the docker source code and found that there are three places where the same error might be output. But I still don't have time to verify it. |
Right. The |
If we can continue to reproduce this problem, then I am very happy to verify and fix it after the holiday (on the docker side) It seems that the most recent CIs have passed. 😂 |
I already sent a fix :-) #974 (comment) -- working on verification. CI failed twice in the past 3 times https://prow.k8s.io/?job=ci-kind-build |
oh: do you mean in |
Yes, on the docker side. I think this user case should be quite a lot. (concurrent mount) If I remember correctly, the last time I saw it was probably in the code related to the file copy.(I just don’t have a computer now, I can’t verify it in more detail. |
so far I've not been able to hit the bug again (testing in #648), but I'm reasonably confident in the solution. |
Then we can keep your solution👍 after three days, my holiday is over, I can do more testing and verification. |
/assign
/priority important-soon
/lifecycle active
docker: Error response from daemon: failed to mkdir /docker-graph/volumes/kind-build-cache/_data/bin: mkdir /docker-graph/volumes/kind-build-cache/_data/bin: file exists.
see:
#961 (comment)
https://storage.googleapis.com/kubernetes-jenkins/pr-logs/pull/kubernetes-sigs_kind/961/pull-kind-build/1185113192976617474/build-log.txt
https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kind-build/1185845621353877504
https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/kubernetes-sigs_kind/648/pull-kind-build/1185680916383535105
The text was updated successfully, but these errors were encountered: