Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

chore: Fix scientific notation in example yamls #2688

Merged
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
description: mnist-estimator-single
hyperparameters:
learning_rate: 1e-3
learning_rate: 1.0e-3
global_batch_size: 64
hidden_layer_1: 2500
hidden_layer_2: 1000
Expand Down
2 changes: 1 addition & 1 deletion e2e_tests/tests/fixtures/mnist_estimator/single.yaml
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
description: mnist-estimator-single
hyperparameters:
learning_rate: 1e-3
learning_rate: 1.0e-3
global_batch_size: 64
hidden_layer_1: 2500
hidden_layer_2: 1000
Expand Down
2 changes: 1 addition & 1 deletion examples/computer_vision/cifar10_pytorch/adaptive.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ hyperparameters:
minval: -5.0
maxval: 1.0
base: 10.0
learning_rate_decay: 1e-6
learning_rate_decay: 1.0e-6
layer1_dropout:
type: double
minval: 0.2
Expand Down
4 changes: 2 additions & 2 deletions examples/computer_vision/cifar10_pytorch/const.yaml
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
name: cifar10_pytorch_const
hyperparameters:
learning_rate: 1e-4
learning_rate_decay: 1e-6
learning_rate: 1.0e-4
learning_rate_decay: 1.0e-6
layer1_dropout: 0.25
layer2_dropout: 0.25
layer3_dropout: 0.5
Expand Down
4 changes: 2 additions & 2 deletions examples/computer_vision/cifar10_pytorch/distributed.yaml
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
name: cifar10_pytorch_distributed
hyperparameters:
learning_rate: 1e-4
learning_rate_decay: 1e-6
learning_rate: 1.0e-4
learning_rate_decay: 1.0e-6
layer1_dropout: 0.25
layer2_dropout: 0.25
layer3_dropout: 0.5
Expand Down
2 changes: 1 addition & 1 deletion examples/computer_vision/cifar10_tf_keras/adaptive.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ hyperparameters:
minval: -5.0
maxval: 1.0
base: 10.0
learning_rate_decay: 1e-6
learning_rate_decay: 1.0e-6
layer1_dropout:
type: double
minval: 0.2
Expand Down
4 changes: 2 additions & 2 deletions examples/computer_vision/cifar10_tf_keras/const.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -2,8 +2,8 @@ name: cifar10_tf_keras_const
data:
url: https://s3-us-west-2.amazonaws.com/determined-ai-datasets/cifar10/cifar-10-python.tar.gz
hyperparameters:
learning_rate: 1e-4
learning_rate_decay: 1e-6
learning_rate: 1.0e-4
learning_rate_decay: 1.0e-6
layer1_dropout: 0.25
layer2_dropout: 0.25
layer3_dropout: 0.5
Expand Down
4 changes: 2 additions & 2 deletions examples/computer_vision/cifar10_tf_keras/distributed.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -2,8 +2,8 @@ name: cifar10_tf_keras_distributed
data:
url: https://s3-us-west-2.amazonaws.com/determined-ai-datasets/cifar10/cifar-10-python.tar.gz
hyperparameters:
learning_rate: 1e-4
learning_rate_decay: 1e-6
learning_rate: 1.0e-4
learning_rate_decay: 1.0e-6
layer1_dropout: 0.25
layer2_dropout: 0.25
layer3_dropout: 0.5
Expand Down
Original file line number Diff line number Diff line change
@@ -1,15 +1,15 @@
name: deformabledetr_coco_fake_data
hyperparameters:
lr: 2e-4
lr: 2.0e-4
lr_backbone_names:
- backbone.0
lr_backbone: 2e-5
lr_backbone: 2.0e-5
lr_linear_proj_names:
- reference_points
- sampling_offsets
lr_linear_proj_mult: 0.1
global_batch_size: 1
weight_decay: 1e-4
weight_decay: 1.0e-4
lr_drop: 40
clip_max_norm: 0.1

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -2,16 +2,16 @@ name: deformabledetr_coco_distributed
hyperparameters:
# These settings match those used in this experiment from the original repo:
# https://github.com/fundamentalvision/Deformable-DETR/blob/main/configs/r50_deformable_detr.sh
lr: 2e-4
lr: 2.0e-4
lr_backbone_names:
- backbone.0
lr_backbone: 2e-5
lr_backbone: 2.0e-5
lr_linear_proj_names:
- reference_points
- sampling_offsets
lr_linear_proj_mult: 0.1
global_batch_size: 32
weight_decay: 1e-4
weight_decay: 1.0e-4
lr_drop: 40
clip_max_norm: 0.1

Expand Down
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
name: deformabledetr_coco_finetune
hyperparameters:
lr: 1e-4
lr: 1.0e-4
lr_backbone_names:
- backbone.0
lr_backbone: 0
Expand All @@ -9,7 +9,7 @@ hyperparameters:
- sampling_offsets
lr_linear_proj_mult: 0.1
global_batch_size: 2
weight_decay: 1e-4
weight_decay: 1.0e-4
lr_drop: 4
clip_max_norm: 0.1

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -2,8 +2,8 @@ name: deformabledetr_coco_adaptive
hyperparameters:
lr:
type: double
minval: 1e-5
maxval: 1e-4
minval: 1.0e-5
maxval: 1.0e-4
lr_backbone_names:
- backbone.0
lr_backbone: 0
Expand All @@ -12,7 +12,7 @@ hyperparameters:
- sampling_offsets
lr_linear_proj_mult: 0.1
global_batch_size: 4
weight_decay: 1e-4
weight_decay: 1.0e-4
lr_drop: 4
clip_max_norm:
type: double
Expand Down
6 changes: 3 additions & 3 deletions examples/computer_vision/detr_coco_pytorch/const_fake.yaml
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
name: detr_coco_fake_data
hyperparameters:
lr: 1e-4
lr_backbone: 1e-5
lr: 1.0e-4
lr_backbone: 1.0e-5
global_batch_size: 2
weight_decay: 1e-4
weight_decay: 1.0e-4
lr_drop: 100
clip_max_norm: 0.1

Expand Down
6 changes: 3 additions & 3 deletions examples/computer_vision/detr_coco_pytorch/distributed.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -2,10 +2,10 @@ name: detr_coco_distributed
hyperparameters:
# These settings match that for the 150 epoch run provided in the original repo:
# https://github.com/facebookresearch/detr
lr: 1e-4
lr_backbone: 1e-5
lr: 1.0e-4
lr_backbone: 1.0e-5
global_batch_size: 16
weight_decay: 1e-4
weight_decay: 1.0e-4
lr_drop: 100
clip_max_norm: 0.1

Expand Down
4 changes: 2 additions & 2 deletions examples/computer_vision/detr_coco_pytorch/finetune.yaml
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
name: detr_coco_finetune
hyperparameters:
lr: 1e-4
lr: 1.0e-4
lr_backbone: 0
global_batch_size: 4
weight_decay: 1e-4
weight_decay: 1.0e-4
lr_drop: 4
clip_max_norm: 0.1

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -3,11 +3,11 @@ hyperparameters:
# We will tune learning rate and gradient clipping.
lr:
type: double
minval: 1e-5
maxval: 1e-4
minval: 1.0e-5
maxval: 1.0e-4
lr_backbone: 0
global_batch_size: 4
weight_decay: 1e-4
weight_decay: 1.0e-4
lr_drop: 4
clip_max_norm:
type: double
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -55,7 +55,7 @@ hyperparameters:
lr_cycle_mul: 1.0
lr_cycle_limit: 1
warmup_lr: 0.0001
min_lr: 1e-5
min_lr: 1.0e-5
start_epoch: None
decay_epochs: 30
warmup_epochs: 5
Expand Down
4 changes: 2 additions & 2 deletions examples/computer_vision/efficientdet_pytorch/const.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@ hyperparameters:
opt: fusedmomentum
opt_eps: 0.001
momentum: 0.9
weight_decay: 4e-05
weight_decay: 4.0e-05
sched: cosine
lr: .03
lr_noise: 0.4 0.9
Expand All @@ -35,7 +35,7 @@ hyperparameters:
lr_cycle_mul: 1.0
lr_cycle_limit: 1
warmup_lr: 0.0001
min_lr: 1e-5
min_lr: 1.0e-5
start_epoch: None
decay_epochs: 30
warmup_epochs: 5
Expand Down
4 changes: 2 additions & 2 deletions examples/computer_vision/efficientdet_pytorch/const_fake.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@ hyperparameters:
opt: fusedmomentum
opt_eps: 0.001
momentum: 0.9
weight_decay: 4e-05
weight_decay: 4.0e-05
sched: cosine
lr: .03
lr_noise: 0.4 0.9
Expand All @@ -35,7 +35,7 @@ hyperparameters:
lr_cycle_mul: 1.0
lr_cycle_limit: 1
warmup_lr: 0.0001
min_lr: 1e-5
min_lr: 1.0e-5
start_epoch: None
decay_epochs: 30
warmup_epochs: 5
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@ hyperparameters:
opt: fusedmomentum
opt_eps: 0.001
momentum: 0.9
weight_decay: 4e-05
weight_decay: 4.0e-05
sched: cosine
lr: 0.06
lr_noise: 0.4 0.9
Expand All @@ -37,7 +37,7 @@ hyperparameters:
lr_cycle_mul: 1.0
lr_cycle_limit: 1
warmup_lr: 0.0001
min_lr: 1e-5
min_lr: 1.0e-5
start_epoch: None
decay_epochs: 30
warmup_epochs: 5
Expand Down
2 changes: 1 addition & 1 deletion examples/computer_vision/iris_tf_keras/adaptive.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ hyperparameters:
minval: -5.0
maxval: 1.0
base: 10.0
learning_rate_decay: 1e-6
learning_rate_decay: 1.0e-6
layer1_dense_size:
type: int
minval: 4
Expand Down
4 changes: 2 additions & 2 deletions examples/computer_vision/iris_tf_keras/const.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,8 +3,8 @@ data:
train_url: http://download.tensorflow.org/data/iris_training.csv
test_url: http://download.tensorflow.org/data/iris_test.csv
hyperparameters:
learning_rate: 1e-4
learning_rate_decay: 1e-6
learning_rate: 1.0e-4
learning_rate_decay: 1.0e-6
layer1_dense_size: 16
global_batch_size: 30
searcher:
Expand Down
4 changes: 2 additions & 2 deletions examples/computer_vision/iris_tf_keras/distributed.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,8 +3,8 @@ data:
train_url: http://download.tensorflow.org/data/iris_training.csv
test_url: http://download.tensorflow.org/data/iris_test.csv
hyperparameters:
learning_rate: 1e-4
learning_rate_decay: 1e-6
learning_rate: 1.0e-4
learning_rate_decay: 1.0e-6
layer1_dense_size: 16
global_batch_size: 30
resources:
Expand Down
2 changes: 1 addition & 1 deletion examples/computer_vision/mnist_estimator/const.yaml
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
name: mnist_estimator_const
hyperparameters:
learning_rate: 1e-3
learning_rate: 1.0e-3
global_batch_size: 64
hidden_layer_1: 2500
hidden_layer_2: 1000
Expand Down
2 changes: 1 addition & 1 deletion examples/computer_vision/mnist_estimator/distributed.yaml
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
name: mnist_estimator_distributed
hyperparameters:
learning_rate: 1e-3
learning_rate: 1.0e-3
global_batch_size: 1024 # per GPU batch size of 64
hidden_layer_1: 2500
hidden_layer_2: 1000
Expand Down
2 changes: 1 addition & 1 deletion examples/computer_vision/mnist_tf_layers/const.yaml
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
name: mnist_tf_core_to_estimator
hyperparameters:
learning_rate: 1e-3
learning_rate: 1.0e-3
global_batch_size: 64
n_filters_1: 10
n_filters_2: 40
Expand Down
4 changes: 2 additions & 2 deletions examples/computer_vision/unets_tf_keras/const.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,8 +4,8 @@ data:
data_file: mobilenet_v2_weights_tf_dim_ordering_tf_kernels_1.0_128_no_top.h5

hyperparameters:
learning_rate: 1e-4
learning_rate_decay: 1e-6
learning_rate: 1.0e-4
learning_rate_decay: 1.0e-6
layer1_dense_size: 16
global_batch_size: 64
OUTPUT_CHANNELS: 3
Expand Down
4 changes: 2 additions & 2 deletions examples/computer_vision/unets_tf_keras/distributed.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,8 +4,8 @@ data:
data_file: mobilenet_v2_weights_tf_dim_ordering_tf_kernels_1.0_128_no_top.h5

hyperparameters:
learning_rate: 1e-4
learning_rate_decay: 1e-6
learning_rate: 1.0e-4
learning_rate_decay: 1.0e-6
layer1_dense_size: 16
global_batch_size: 512 # per slot batch size = 64
OUTPUT_CHANNELS: 3
Expand Down
2 changes: 1 addition & 1 deletion examples/features/data_layer_mnist_estimator/const.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@ name: data_layer_mnist_estimator_const
data:
skip_checkpointing_input: true
hyperparameters:
learning_rate: 1e-3
learning_rate: 1.0e-3
global_batch_size: 16
hidden_layer_1: 2500
hidden_layer_2: 1000
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@ name: data_layer_mnist_estimator_const
data:
skip_checkpointing_input: true
hyperparameters:
learning_rate: 1e-3
learning_rate: 1.0e-3
global_batch_size: 128
hidden_layer_1: 2500
hidden_layer_2: 1000
Expand Down
4 changes: 2 additions & 2 deletions examples/features/data_layer_mnist_tf_keras/const.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -2,8 +2,8 @@ name: data_layer_mnist_tf_keras_const
data:
url: https://s3-us-west-2.amazonaws.com/determined-ai-datasets/cifar10/cifar-10-python.tar.gz
hyperparameters:
learning_rate: 1e-4
learning_rate_decay: 1e-6
learning_rate: 1.0e-4
learning_rate_decay: 1.0e-6
layer1_dropout: 0.25
layer2_dropout: 0.25
layer3_dropout: 0.5
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@ min_validation_period:
hyperparameters:
learning_rate: 0.025
momentum: 0.9
weight_decay: 3e-4
weight_decay: 3.0e-4
train_epochs: 300
global_batch_size: 96
init_channels: 36
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@ hyperparameters:
use_constraints: true
learning_rate: 0.025
momentum: 0.9
weight_decay: 3e-4
weight_decay: 3.0e-4
train_epochs: 300
global_batch_size: 96
init_channels: 36
Expand Down
Loading