Skip to content

Commit 783b57e

Browse files
pstjohncamirr-nv
authored andcommitted
Add pre commit to verify test status (#841)
Previously if pre-commit failed, it would cause `run-tests` to be skipped, which would then mean that `verify-tests-status` would give the PR the green light. That's obviously a problem, so we need to make sure all these tests are added to the verify-tests-status check. Ideally we could put some more logic in that if block to check if we're in a merge queue, whether tests were intentionally skipped, etc. --------- Signed-off-by: Peter St. John <pstjohn@nvidia.com> Signed-off-by: Ubuntu <camirr@nvidia.com>
1 parent 2fa3b86 commit 783b57e

File tree

2 files changed

+5
-2
lines changed

2 files changed

+5
-2
lines changed

.github/workflows/unit-tests.yml

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -207,6 +207,9 @@ jobs:
207207
# This steps checks the status of all test jobs and fails if any of them failed or were cancelled.
208208
# It is a workaround for the lack of a built-in feature to finalize a pipeline by checking the status of multiple jobs
209209
needs: # List all your run-*-test jobs
210+
- pre-commit
211+
- get-pr-labels
212+
- build-bionemo-image
210213
- run-tests
211214
- run-slow-tests
212215
- run-notebooks-docs

docs/docs/models/geneformer.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -150,11 +150,11 @@ NVIDIA believes Trustworthy AI is a shared responsibility and we have establishe
150150

151151
### geneformer-10M
152152
<!-- WandB Logs: https://wandb.ai/clara-discovery/Geneformer-pretraining-jsjconfigs/runs/i8LWOctg?nw=nwuserjomitchell -->
153-
Training was performed on 8 servers with 8 A100 GPUs each for a total of 81485 steps using the CELLxGENE split with a per-gpu micro batch size 32 and global batch size of 2048. Training took a total of 4 days, 8 hours of wallclock time. As can be seen in the following images, training and validation curves both decreased fairly smoothly throughout the course of training.
153+
Training was performed on 8 servers with 8 A100 GPUs each for a total of 81485 steps using the CELLxGENE split with a per-gpu micro batch size 32 and global batch size of 2048. Training took a total of 4 days, 8 hours of wallclock time. As can be seen in the following images, training and validation curves both decreased fairly smoothly throughout the course of training.
154154

155155
![Training Loss Geneformer 10M](../assets/images/geneformer/geneformer_10m_training_loss.png)
156156
![Validation Loss Geneformer 10M](../assets/images/geneformer/geneformer_10m_val_loss.png)
157-
157+
158158

159159

160160
### geneformer-106M

0 commit comments

Comments
 (0)