|
48 | 48 | add_arg('--epoch-size', default=72, type=int, help='Number of batches trained in an epoch.') |
49 | 49 | add_arg('--save-every', default=10, type=int, help='Save generator after every training epoch.') |
50 | 50 | add_arg('--batch-shape', default=192, type=int, help='Resolution of images in training batch.') |
51 | | -add_arg('--batch-size', default=10, type=int, help='Number of images per training batch.') |
| 51 | +add_arg('--batch-size', default=15, type=int, help='Number of images per training batch.') |
52 | 52 | add_arg('--buffer-size', default=1500, type=int, help='Total image fragments kept in cache.') |
53 | 53 | add_arg('--buffer-similar', default=5, type=int, help='Fragments cached for each image loaded.') |
54 | | -add_arg('--learning-rate', default=5E-4, type=float, help='Parameter for the ADAM optimizer.') |
55 | | -add_arg('--learning-period', default=100, type=int, help='How often to decay the learning rate.') |
| 54 | +add_arg('--learning-rate', default=1E-4, type=float, help='Parameter for the ADAM optimizer.') |
| 55 | +add_arg('--learning-period', default=50, type=int, help='How often to decay the learning rate.') |
56 | 56 | add_arg('--learning-decay', default=0.5, type=float, help='How much to decay the learning rate.') |
57 | 57 | add_arg('--generator-upscale', default=2, type=int, help='Steps of 2x up-sampling as post-process.') |
58 | 58 | add_arg('--generator-downscale',default=0, type=int, help='Steps of 2x down-sampling as preprocess.') |
|
0 commit comments