-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comparison to DAT? #3
Comments
Hi, Thank you for your detailed testing. The latency of RGT and DAT is close, which aligns with the calculated results of Params and FLOPs. However, RGT performs better, for instance, on the Urban100 dataset, with the following specific comparisons:
This is because, unlike DAT, which employs channel attention to achieve linear complexity, RGT adopts linear global spatial attention, which is more suited for SR tasks. |
Sound good, thank you for your work :) I was able to train an RGT and an RGT-S model. Results look good. My latest RGT-S one can be found here. Its called '4xRealWebPhoto_v2_rgt_s' since its idea was to upscale downloaded photos from the web, so i modeled the degradations of the dataset to include scaling and compression and rescaling and recompression (by a service provider like the user uploading to social media or something, then someone else downloading and re-uploading again. Also added realistic noise and some slight lens blur). |
@Phhofm Hi. I was wondering if you could create a tutorial on how to further train models. I have an RTX 3090 graphics card with 24GB VRAM and 32GB of RAM, and I'd like to train a model for the first time. By the way, I've been using open-source projects from GitHub for 4-5 years now, so I believe with a little help I can get the hang of it. Thanks a lot! |
Hey, you could definetly get into training your own upscaling models. |
@Phhofm Would you mind updating the link to the Google Drive folder with new models in your "models" GitHub repository? I've successfully used some of them in Hybrid, a video editing program. Thanks! |
Hey, thank you for your work. I had just a quick question.
I didnt look (or understand) into the technical details too much, I am writing simply from the practical/applying side (tried training some models).
Performance wise on an application standpoint RGT feels fairly similar to DAT, meaning training and inference speed wise and output quality wise, just to me personally.
I also just recently did an inference speed test, and RGT got very similar speeds to DAT (and RGT-S to DAT-S)
I simply wanted to ask on the theoretical side, if there maybe would be a usecase where DAT would be preferable (/should perform better) over RGT or RGT preferable over DAT?
Ah PS here some outputs / examples of an RGT model I recently trained:
Slowpoke Pics 6 Examples
---- mentioned inference speed test:
4x Inference speed test, neosr testscripts, 50 256x256 images as input, Ubuntu (Budgie) 23.10, GeForce RTX 3060 Lite Hash Rate, AMD Ryzen™ 5 3600 × 12
Sorted network option according to fastest run out of 3 runs each (first started with more, so compact and cugan had more, but then hit dat and from then on 3 each):
Compact: 1.90s, 26.35fps
SPAN: 2.33s, 21.44fps
SAFMN: 2.51s, 19.89fps
DITN: 4.26s, 11.72fps
CUGAN: 4.45s, 11.22fps
OmniSR: 8.90s, 5.62fps
SAFMN-L: 9.87s, 5.07fps
CRAFT: 11.26s, 4.44fps
DCTLSA: 11.53s, 4.43fps
SwinIR-S: 14.18s, 3.53fps
SRFormer-light: 16.28s, 3.07fps
ESRGAN: 22.51s, 2.22fps
SwinIR-M: 46.46s, 1.08fps
HAT-S: 71.37s, 0.70fps
RGT_S: 74.83s, 0.67fps
DAT-S: 74.96s, 0.67fps
SRFormer-M: 79.02s, 0.63fps
DAT2: 81.90s, 0.61fps
HAT-M: 90.19s, 0.55fps
RGT: 96.07s, 0.52fps
DAT: 97.08s, 0.52fps
HAT-L: 177.75s, 0.28fps
The text was updated successfully, but these errors were encountered: