-
Notifications
You must be signed in to change notification settings - Fork 350
Add more tests to cover containerd/driver container k8s deployments #1139
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
9dfb377 to
d30b231
Compare
| timeout 180 bash -x ./workloads/jenkins/scripts/test-dashboard.sh | ||
| ''' | ||
|
|
||
| echo "Start new virtual environment pre-Slurm checks" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do we need to explicitly tear down the VMs before we do this?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Tearing them down is part of this script.
3c4471d to
efcf7aa
Compare
efcf7aa to
dbf4c87
Compare
ajdecon
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM. Confirmed that tests passed in nightly builds.
We now have the following configurations to test.
This change covers these all in the nightlies and it will probably take around 4 hours end to end.
I tried to mix-and-match the actual tests being run across them so that we have some confidence without filling out a dense test matrics. Kubeflow is tested once on GPU Operator and once on device plugin.
I skipped the local-registry tests on containerd installs, but kept it in for docker installs. This may or may not work given the recent changes.
I made sure that the monitoring stack is tested with at least one configuration of device plugin, docker, containerd, and driver-container configurations.