-
Notifications
You must be signed in to change notification settings - Fork 98
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Training a smaller edge model #11
Comments
Hi Rishabh, I was also thinking about this kind of thing to help the deployment on edge devices. However, I need to pay for renting GPUs with my own money for training. Now that here is the question, I'll try to do it in recent days. The easiest way to train a smaller model is, of course, to use Swin-tiny as the backbone network. |
Hi, I've added choices of more lightweight backbones in the codes. You can now try lightweight models like swin_v1_tiny, swin_v1_small, pvt_v2_b0, pvt_v2_b1, and pvt_v2_b2 as the backbone. But training of models with them is temporarily suspended due to the situation I mentioned above.
|
can you also share the param numbers ? also can we you do a quality check when using 8 bit or 16 bit ? |
You can try |
Can you share the param count . For the swin tiny backbone ? |
You can choose the |
Hi, I finished the 500-epoch training on a BiRefNet with +---------+----------------------------+-------+-----------+------+----------+--------+------+-------+--------+-------+-------+ |
Ohh great thanks 😃 |
Feel free to reopen it if there are any relevant questions. |
Hi, @rishabh063 , I've updated a BiRefNet for general segmentation with swin_v1_tiny as the backbone for the edge device. The well-trained model has been uploaded to my Google Drive. Check the Meanwhile, check the update in inference.py. Set the torch.set_float32_matmul_precision to 'high' can increase the FPS of the large version on A100 from 5 to 12 with ~0 performance downgrade (Because I set it to 'high' during training). Good luck with the smaller and faster BiRefNet with ~0 degradation. |
hey thanks , i will try it out in some time . swamped with work btw any idea on that SOD bounding box ? |
That's still in progress, too much training in the queue. I'll reply to you here once it's done. |
What backbone are you thinking to use there ? I will also try training that . New to model training .would love all your help @ZhengPeng7 ( forgot to mention yesterday) |
I'll try both swin_v1_large and swin_v1_tiny for the best performance and efficiency, respectively. If still too heavy, I'll try pvt_v2_b1. |
okay i will be using u2netp without any backbone , lets see if i can get anything |
Hey @ZhengPeng7 hope you are well . I am thinking to training a smaller model to do downstream task. Any insights where i can shrink param count ?
The text was updated successfully, but these errors were encountered: