diff --git a/MODEL_ZOO.md b/MODEL_ZOO.md index d83c588..0274d2b 100644 --- a/MODEL_ZOO.md +++ b/MODEL_ZOO.md @@ -58,8 +58,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 10 2.8 31.1 -000000000 -model +160905981 +model @@ -71,8 +71,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 15 3.9 27.3 -000000000 -model +160905967 +model @@ -84,8 +84,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 17 4.4 25.9 -000000000 -model +160906442 +model @@ -97,8 +97,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 21 5.7 24.8 -000000000 -model +160906036 +model @@ -110,8 +110,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 33 8.7 23.0 -000000000 -model +160990626 +model @@ -123,8 +123,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 57 14.3 21.7 -000000000 -model +160906139 +model @@ -136,8 +136,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 69 17.1 21.4 -000000000 -model +160906383 +model @@ -149,8 +149,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 92 23.5 20.8 -000000000 -model +161116590 +model @@ -162,8 +162,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 94 22.6 20.7 -000000000 -model +161107726 +model @@ -175,8 +175,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 137 32.9 20.3 -000000000 -model +160906020 +model @@ -188,8 +188,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 168 39.7 20.0 -000000000 -model +158460855 +model @@ -202,7 +202,7 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 76.9 19.5 000000000 -model +model @@ -233,8 +233,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 11 3.1 29.6 -000000000 -model +176245422 +model @@ -246,8 +246,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 19 5.1 25.9 -000000000 -model +160906449 +model @@ -259,8 +259,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 19 5.2 24.5 -000000000 -model +160981443 +model @@ -272,8 +272,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 22 6.0 23.7 -000000000 -model +160906567 +model @@ -285,8 +285,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 39 10.1 22.0 -000000000 -model +160906681 +model @@ -298,8 +298,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 67 16.5 21.0 -000000000 -model +160906834 +model @@ -311,8 +311,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 68 16.8 20.6 -000000000 -model +160906838 +model @@ -324,8 +324,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 104 26.1 20.1 -000000000 -model +160907112 +model @@ -337,8 +337,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 113 28.1 20.1 -000000000 -model +161160905 +model @@ -350,8 +350,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 150 36.0 19.7 -000000000 -model +160907100 +model @@ -363,8 +363,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 189 45.6 19.6 -000000000 -model +161303400 +model @@ -376,8 +376,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 319 76.0 19.0 -000000000 -model +161277763 +model @@ -408,8 +408,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 53 12.2 23.2 -000000000 -model +161235311 +model @@ -421,8 +421,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 90 20.4 21.4 -000000000 -model +161167170 +model @@ -434,8 +434,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 130 29.2 20.9 -000000000 -model +161167467 +model @@ -466,8 +466,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 78 18.0 21.9 -000000000 -model +161167411 +model @@ -479,8 +479,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 137 31.8 20.7 -000000000 -model +161167590 +model @@ -492,8 +492,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 197 45.7 20.4 -000000000 -model +162471172 +model @@ -524,8 +524,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 34 11.7 24.9 -000000000 -model +161305613 +model @@ -537,8 +537,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 52 15.6 24.1 -000000000 -model +161304979 +model @@ -550,8 +550,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 68 18.4 23.4 -000000000 -model +161305015 +model @@ -563,8 +563,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 114 32.1 22.5 -000000000 -model +161305060 +model @@ -576,8 +576,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 240 65.1 21.2 -000000000 -model +161305098 +model @@ -589,8 +589,8 @@ For 8 GPU training, we apply 5 epoch gradual warmup, following the [ImageNet in 504 135.1 21.5 -000000000 -model +161305138 +model diff --git a/configs/dds_baselines/effnet/EN-B0_dds_8gpu.yaml b/configs/dds_baselines/effnet/EN-B0_dds_8gpu.yaml index 2d458be..4f87ee4 100644 --- a/configs/dds_baselines/effnet/EN-B0_dds_8gpu.yaml +++ b/configs/dds_baselines/effnet/EN-B0_dds_8gpu.yaml @@ -8,6 +8,7 @@ EN: EXP_RATIOS: [1, 6, 6, 6, 6, 6, 6] KERNELS: [3, 3, 5, 3, 5, 5, 3] HEAD_W: 1280 + STEM_W: 32 OPTIM: LR_POLICY: cos BASE_LR: 0.4 diff --git a/configs/dds_baselines/resnext/X-50-32x4d_dds_8gpu.yaml b/configs/dds_baselines/resnext/X-50-32x4d_dds_8gpu.yaml index bc98c2b..0b6f1ed 100644 --- a/configs/dds_baselines/resnext/X-50-32x4d_dds_8gpu.yaml +++ b/configs/dds_baselines/resnext/X-50-32x4d_dds_8gpu.yaml @@ -9,7 +9,7 @@ ANYNET: DEPTHS: [3, 4, 6, 3] WIDTHS: [256, 512, 1024, 2048] BOT_MULS: [0.5, 0.5, 0.5, 0.5] - GROUP_WS: [4, 8, 16, 32] + GROUP_WS: [64, 8, 16, 32] OPTIM: LR_POLICY: cos BASE_LR: 0.2