Hello, I encountered some issues when reproducing your model. When evaluating the public model weights (ADE20K, UperNet-R50) using Mask2Former, the mIoU was 43.46, which matches the revised metrics in your paper.However, the highest metric I achieved during reproduction was 41.18 (on RTX 4090 D-24G2 with train_batch_size=1 and gradient_accumulation_steps=4). I have tried using different devices such as A100-80G2, H20-96G4, and A100-40G1, but none of them achieved higher metrics.
Given the limited resources at our school, Can you provide me with some guidance on modifying training parameters to achieve the desired results when resources are limited?
Hello, I encountered some issues when reproducing your model. When evaluating the public model weights (ADE20K, UperNet-R50) using Mask2Former, the mIoU was 43.46, which matches the revised metrics in your paper.However, the highest metric I achieved during reproduction was 41.18 (on RTX 4090 D-24G2 with train_batch_size=1 and gradient_accumulation_steps=4). I have tried using different devices such as A100-80G2, H20-96G4, and A100-40G1, but none of them achieved higher metrics.
Given the limited resources at our school, Can you provide me with some guidance on modifying training parameters to achieve the desired results when resources are limited?