Skip to content

Commit 9317709

Browse files
Merge pull request #32 from PeterCalifano/dev_improve_debug_augmentations
Extend image augmentation module and definitely fix device bug
2 parents ff44b93 + 8086d67 commit 9317709

File tree

11 files changed

+1353
-705
lines changed

11 files changed

+1353
-705
lines changed

TODO

Lines changed: 11 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -32,16 +32,25 @@
3232
- [x] Add automatic mixed precision support from torch
3333
- [x] Add optional validation check in Augmentation helper to ensure all inputs are valid ones with remedy action
3434
- [x] Improve handling of training errors (stop and kill run if nan or Inf)
35+
- [x] Implement prototype usage of AMP for mixed precision training
3536

36-
- Patches of v0.2.X
37+
- Minor versions and patches of v0.2.X
3738
- [x] Major fixes in v0.2.1 (bugs introduced in v0.2.0)
39+
- [x] Fix bugs in augmentation module (validation module)
40+
- [ ] [MAJOR] Restructure and expand augmentations module to correctly handle images and 1D vectors jointly
41+
- [ ] Implement 1d vector error models (from selected distributions)
42+
- [ ] Change random apply and structure of augs module for images (split based on type)
43+
- [x] Implement "binarize" augmentation for images (EDIT: soft-binarize in the final version)
44+
- [ ] Review and unit test selection criteria implementation
45+
- [ ] Extend augs-trainer integration to allow "discard" option in validation module
46+
- [x] Implement tailoring of RandomAffine to make it border crossing aware
3847

3948
- [ ] v0.3.X
4049
- [x] Implement custom adaptive pooling layers for ONNx static export
4150
- [x] Implement tests for *onnxability* of models provided by model_building module of PTAF
4251
- [ ] Modify implementation of trainer to use static or cls methods instead of instance methods for increased flexibility
4352
- [ ] Add capability to trainer/new class helper: transfer learning from checkpoint instead of replacing model!
44-
- [ ] Fix unit tests
53+
- [ ] Fix all unit tests for up-to-date modules
4554
- [ ] Implement export method (traced, onnx, model pth) in ModelTrainingManager
4655
- [ ] Add abstract meta class for loss functions to enforce interface
4756
- [ ] Review checkpoint resuming code (ensure the checkpoint is loaded correctly) and add evaluation before training (baseline score)
@@ -50,7 +59,6 @@
5059
- [ ] Modify DataLoaderIndex to accept datasets directly and a combination of datasets and dataloaders. If dataset is input, use default specifications for dataloader
5160
- [ ] Implement new training mode: SWA_MODE
5261
- [ ] Add pruning strategy in NORMAL MODE using "delta loss" over patience interval to check for pruning (like OPTUNA)
53-
- [ ] Implement prototype usage of AMP for mixed precision training
5462
- [ ] Add configurable dataclasses from template yml files for training manager and extensible prototype (see pySR configuration in nav-frontend)
5563
- [ ] Add pySR conveniency module
5664
- [ ] Explore Hydra for configuration management from yml files

lib/CommManager4MATLAB

0 commit comments

Comments
 (0)