|
32 | 32 | - [x] Add automatic mixed precision support from torch |
33 | 33 | - [x] Add optional validation check in Augmentation helper to ensure all inputs are valid ones with remedy action |
34 | 34 | - [x] Improve handling of training errors (stop and kill run if nan or Inf) |
| 35 | + - [x] Implement prototype usage of AMP for mixed precision training |
35 | 36 |
|
36 | | -- Patches of v0.2.X |
| 37 | +- Minor versions and patches of v0.2.X |
37 | 38 | - [x] Major fixes in v0.2.1 (bugs introduced in v0.2.0) |
| 39 | + - [x] Fix bugs in augmentation module (validation module) |
| 40 | + - [ ] [MAJOR] Restructure and expand augmentations module to correctly handle images and 1D vectors jointly |
| 41 | + - [ ] Implement 1d vector error models (from selected distributions) |
| 42 | + - [ ] Change random apply and structure of augs module for images (split based on type) |
| 43 | + - [x] Implement "binarize" augmentation for images (EDIT: soft-binarize in the final version) |
| 44 | + - [ ] Review and unit test selection criteria implementation |
| 45 | + - [ ] Extend augs-trainer integration to allow "discard" option in validation module |
| 46 | + - [x] Implement tailoring of RandomAffine to make it border crossing aware |
38 | 47 |
|
39 | 48 | - [ ] v0.3.X |
40 | 49 | - [x] Implement custom adaptive pooling layers for ONNx static export |
41 | 50 | - [x] Implement tests for *onnxability* of models provided by model_building module of PTAF |
42 | 51 | - [ ] Modify implementation of trainer to use static or cls methods instead of instance methods for increased flexibility |
43 | 52 | - [ ] Add capability to trainer/new class helper: transfer learning from checkpoint instead of replacing model! |
44 | | - - [ ] Fix unit tests |
| 53 | + - [ ] Fix all unit tests for up-to-date modules |
45 | 54 | - [ ] Implement export method (traced, onnx, model pth) in ModelTrainingManager |
46 | 55 | - [ ] Add abstract meta class for loss functions to enforce interface |
47 | 56 | - [ ] Review checkpoint resuming code (ensure the checkpoint is loaded correctly) and add evaluation before training (baseline score) |
|
50 | 59 | - [ ] Modify DataLoaderIndex to accept datasets directly and a combination of datasets and dataloaders. If dataset is input, use default specifications for dataloader |
51 | 60 | - [ ] Implement new training mode: SWA_MODE |
52 | 61 | - [ ] Add pruning strategy in NORMAL MODE using "delta loss" over patience interval to check for pruning (like OPTUNA) |
53 | | - - [ ] Implement prototype usage of AMP for mixed precision training |
54 | 62 | - [ ] Add configurable dataclasses from template yml files for training manager and extensible prototype (see pySR configuration in nav-frontend) |
55 | 63 | - [ ] Add pySR conveniency module |
56 | 64 | - [ ] Explore Hydra for configuration management from yml files |
|
0 commit comments