From a5dc1d59240da6f537f4ebc5c698de544089ca69 Mon Sep 17 00:00:00 2001 From: Chay Ryali Date: Wed, 11 Dec 2024 07:07:23 +0000 Subject: [PATCH] update training config as well to be consistent --- .../sam2.1_training/sam2.1_hiera_b+_MOSE_finetune.yaml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/sam2/configs/sam2.1_training/sam2.1_hiera_b+_MOSE_finetune.yaml b/sam2/configs/sam2.1_training/sam2.1_hiera_b+_MOSE_finetune.yaml index 2046791..9b6faa7 100644 --- a/sam2/configs/sam2.1_training/sam2.1_hiera_b+_MOSE_finetune.yaml +++ b/sam2/configs/sam2.1_training/sam2.1_hiera_b+_MOSE_finetune.yaml @@ -97,7 +97,7 @@ trainer: self_attention: _target_: sam2.modeling.sam.transformer.RoPEAttention rope_theta: 10000.0 - feat_sizes: [32, 32] + feat_sizes: [64, 64] embedding_dim: 256 num_heads: 1 downsample_rate: 1 @@ -108,7 +108,7 @@ trainer: cross_attention: _target_: sam2.modeling.sam.transformer.RoPEAttention rope_theta: 10000.0 - feat_sizes: [32, 32] + feat_sizes: [64, 64] rope_k_repeat: True embedding_dim: 256 num_heads: 1