Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
16 changes: 16 additions & 0 deletions data/architectures.json
Original file line number Diff line number Diff line change
Expand Up @@ -145,6 +145,14 @@
"pytorch"
]
},
"rcan": {
"name": "RCAN",
"input": "image",
"compatiblePlatforms": [
"pytorch",
"onnx"
]
},
"real-cugan": {
"name": "Real-CUGAN",
"input": "image",
Expand Down Expand Up @@ -236,5 +244,13 @@
"pytorch",
"onnx"
]
},
"tscunet": {
"name": "TSCUNet",
"input": "video",
"compatiblePlatforms": [
"pytorch",
"onnx"
]
}
}
31 changes: 30 additions & 1 deletion data/collections.json
Original file line number Diff line number Diff line change
Expand Up @@ -10,6 +10,24 @@
"2x-AnimeSharpV2-MoSR-Soft"
]
},
"c-animesharpv3": {
"name": "AnimeSharpV3",
"author": "kim2091",
"description": "",
"models": [
"2x-AnimeSharpV3",
"2x-AnimeSharpV3-RCAN"
]
},
"c-animesharpv4": {
"name": "AnimeSharpV4",
"author": "kim2091",
"description": "",
"models": [
"2x-AnimeSharpV4",
"2x-AnimeSharpV4-Fast-RCAN-PU"
]
},
"c-normal-map-upscaling": {
"name": "Normal Map Upscaling",
"description": "This collection contain my RG0 normal map upscaling models.\n\nAll models here are for upscaling *tangent-space* normal maps in RG0 format. RG0 means that the B channel is set to 0. These models will work not correctly if you give them images with non-zero B channel, so you either have to zero the B channel manually or use tool like chaiNNer to do it.\n\n## DDS Compression\n\nI made 3 versions: \n- Normal RG0 is for uncompressed normal map textures. Since it hasn't been trained on compression artifacts, it's highly sensitive to quantization artifacts and noise.\n- Normal RG0 BC1 is for BC1-compressed DDS normal map textures.\n- Normal RG0 BC7 is for BC7-compressed DDS normal map textures. This model sometimes produces images that aren't as sharp. In those cases, you can try the BC1 version to see whether it gives better results.",
Expand All @@ -30,7 +48,18 @@
"4x-PBRify-UpscalerSPANV4",
"4x-PBRify-UpscalerSIR-M-V2",
"4x-PBRify-UpscalerDAT2-V1",
"4x-PBRify-RPLKSRd-V3"
"4x-PBRify-RPLKSRd-V3",
"4x-PBRify-UpscalerV4"
],
"author": "kim2091"
},
"c-gameup": {
"name": "GameUp",
"description": "GameUp is a set of video upscaling models intended for upscaling and restoring video game footage",
"models": [
"2x-GameUpV2-TSCUNet",
"2x-GameUpV2-TSCUNet-Small",
"2x-GameUp-TSCUNet"
],
"author": "kim2091"
}
Expand Down
53 changes: 53 additions & 0 deletions data/models/1x-BroadcastToStudio-Compact.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,53 @@
{
"name": "BroadcastToStudio Compact",
"author": "kim2091",
"license": "CC-BY-NC-SA-4.0",
"tags": [
"cartoon",
"compression-removal",
"deblur",
"restoration"
],
"description": "Purpose: Cartoons\n\nThis is a simple retrain of SaurusX's 1x_BroadcastToStudioLite_485k model from a couple years ago. This one is trained on compact, actually has less artifacts, and is significantly faster.\n\n__Comparisons:__ <https://slow.pics/c/oGwHyYym>\n![1733071729 5752351](https://github.com/user-attachments/assets/f247b870-d49e-4c37-ad39-c4565efe164d)",
"date": "2024-12-01",
"architecture": "compact",
"size": [
"64nf",
"16nc"
],
"scale": 1,
"inputChannels": 3,
"outputChannels": 3,
"resources": [
{
"platform": "pytorch",
"type": "pth",
"size": 2400484,
"sha256": "f4876edc5f12783395c444bbb1d2f1bc304ed69b2cc82409dff9d6dad6fbd596",
"urls": [
"https://github.com/Kim2091/Kim2091-Models/releases/download/1x-BroadcastToStudio_Compact/1x-BroadcastToStudio_Compact.pth"
]
},
{
"platform": "onnx",
"type": "onnx",
"size": 1200682,
"sha256": "52836c782140058bcc695e90102c3ef54961ebab2c12e66298eaba25d42570bc",
"urls": [
"https://github.com/Kim2091/Kim2091-Models/releases/download/1x-BroadcastToStudio_Compact/1x-BroadcastToStudio_Compact-fp16.onnx"
]
}
],
"trainingIterations": 25000,
"trainingBatchSize": 8,
"dataset": "BroadcastToStudio",
"datasetSize": 6000,
"pretrainedModelG": "1x-SwatKats-Compact",
"images": [
{
"type": "paired",
"LR": "https://i.slow.pics/DBrq6k0g.webp",
"SR": "https://i.slow.pics/XJmeT86k.webp"
}
]
}
41 changes: 41 additions & 0 deletions data/models/1x-SuperScale-Alt-RPLKSR-S.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,41 @@
{
"name": "SuperScale_Alt_RPLKSR_S",
"author": "kim2091",
"license": "CC-BY-NC-SA-4.0",
"tags": [
"anti-aliasing",
"game-screenshots",
"photo",
"restoration"
],
"description": "Purpose: Anti-aliasing, Restoration\n\nI was bored, so I did this. This model uses DPID as the scaling algorithm for the HRs. The original images were 8k or 12k. It's significantly sharper than Box/Area scaling, yet does a great job with aliasing. This allows for a very sharp model with minimal artifacts, even on the SPAN version.\n\nThe main model is trained on 12k images captured with Nvidia Ansel. It took about 2 days capturing manual 4k and 12k pairs for this model. The 4k captures were used as the LR, the 12k captures were resized to 4k with DPID with randomized lambda values, then trained on as HRs. \n\nThe Alt model is trained exclusively on 8k images from my 8k dataset, resized to 4k with dpid. This provides a clearer result with less noise, but it doesn't handle long edges well at all.\n\nThanks to CF2lter for advice on preparing the dataset, and umzi2 for creating the [rust version of DPID](<https://github.com/umzi2/pepedpid>). \n\n**Showcase:** https://slow.pics/c/TCyqje9K\n![Animation (2)](https://github.com/user-attachments/assets/fb9f010a-a4e1-4537-8b23-9a69974011c6)",
"date": "2025-05-05",
"architecture": "realplksr",
"size": [
"Small",
"Tiny"
],
"scale": 1,
"inputChannels": 3,
"outputChannels": 3,
"resources": [
{
"platform": "pytorch",
"type": "safetensors",
"size": 1195070,
"sha256": "f5844dd72922a6579cf73c44de5cf35f0d700bd407982ae8d03bcdf720924425",
"urls": [
"https://github.com/Kim2091/Kim2091-Models/releases/download/1x-SuperScale/1x-SuperScale_Alt_RPLKSR_S.safetensors"
]
}
],
"dataset": "8k Dataset V3, Custom Ansel dataset",
"images": [
{
"type": "paired",
"caption": "1",
"LR": "https://i.slow.pics/543aXqwG.webp",
"SR": "https://i.slow.pics/o3fwZmcB.webp"
}
]
}
38 changes: 38 additions & 0 deletions data/models/1x-SuperScale-RPLKSR-S.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,38 @@
{
"name": "SuperScale_RPLKSR_S",
"author": "kim2091",
"license": "CC-BY-NC-SA-4.0",
"tags": [
"anti-aliasing",
"game-screenshots",
"photo",
"restoration"
],
"description": "Purpose: Anti-aliasing, Restoration\n\nI was bored, so I did this. This model uses DPID as the scaling algorithm for the HRs. The original images were 8k or 12k. It's significantly sharper than Box/Area scaling, yet does a great job with aliasing. This allows for a very sharp model with minimal artifacts, even on the SPAN version.\n\nThe main model is trained on 12k images captured with Nvidia Ansel. It took about 2 days capturing manual 4k and 12k pairs for this model. The 4k captures were used as the LR, the 12k captures were resized to 4k with DPID with randomized lambda values, then trained on as HRs. \n\nThe Alt model is trained exclusively on 8k images from my 8k dataset, resized to 4k with dpid. This provides a clearer result with less noise, but it doesn't handle long edges well at all.\n\nThanks to CF2lter for advice on preparing the dataset, and umzi2 for creating the [rust version of DPID](<https://github.com/umzi2/pepedpid>). \n\n**Showcase:** https://slow.pics/c/TCyqje9K\n![Animation (2)](https://github.com/user-attachments/assets/fb9f010a-a4e1-4537-8b23-9a69974011c6)",
"date": "2025-05-05",
"architecture": "realplksr",
"size": null,
"scale": 1,
"inputChannels": 3,
"outputChannels": 3,
"resources": [
{
"platform": "pytorch",
"type": "safetensors",
"size": 1195070,
"sha256": "f5844dd72922a6579cf73c44de5cf35f0d700bd407982ae8d03bcdf720924425",
"urls": [
"https://github.com/Kim2091/Kim2091-Models/releases/download/1x-SuperScale/1x-SuperScale_RPLKSR_S.safetensors"
]
}
],
"dataset": "8k Dataset V3, Custom Ansel dataset",
"images": [
{
"type": "paired",
"caption": "1",
"LR": "https://i.slow.pics/543aXqwG.webp",
"SR": "https://i.slow.pics/V1xP5Zh5.webp"
}
]
}
38 changes: 38 additions & 0 deletions data/models/1x-SuperScale.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,38 @@
{
"name": "SuperScale",
"author": "kim2091",
"license": "CC-BY-NC-SA-4.0",
"tags": [
"anti-aliasing",
"game-screenshots",
"photo",
"restoration"
],
"description": "Purpose: Anti-aliasing, Restoration\n\nI was bored, so I did this. This model uses DPID as the scaling algorithm for the HRs. The original images were 8k or 12k. It's significantly sharper than Box/Area scaling, yet does a great job with aliasing. This allows for a very sharp model with minimal artifacts, even on the SPAN version.\n\nThe main model is trained on 12k images captured with Nvidia Ansel. It took about 2 days capturing manual 4k and 12k pairs for this model. The 4k captures were used as the LR, the 12k captures were resized to 4k with DPID with randomized lambda values, then trained on as HRs. \n\nThe Alt model is trained exclusively on 8k images from my 8k dataset, resized to 4k with dpid. This provides a clearer result with less noise, but it doesn't handle long edges well at all.\n\nThanks to CF2lter for advice on preparing the dataset, and umzi2 for creating the [rust version of DPID](<https://github.com/umzi2/pepedpid>). \n\n**Showcase:** https://slow.pics/c/TCyqje9K\n![Animation (2)](https://github.com/user-attachments/assets/fb9f010a-a4e1-4537-8b23-9a69974011c6)",
"date": "2025-05-05",
"architecture": "span",
"size": null,
"scale": 1,
"inputChannels": 3,
"outputChannels": 3,
"resources": [
{
"platform": "pytorch",
"type": "safetensors",
"size": 1195070,
"sha256": "f5844dd72922a6579cf73c44de5cf35f0d700bd407982ae8d03bcdf720924425",
"urls": [
"https://github.com/Kim2091/Kim2091-Models/releases/download/1x-SuperScale/1x-SuperScale_SPAN.safetensors"
]
}
],
"dataset": "8k Dataset V3, Custom Ansel dataset",
"images": [
{
"type": "paired",
"caption": "1",
"LR": "https://i.slow.pics/543aXqwG.webp",
"SR": "https://i.slow.pics/pKA297Fp.webp"
}
]
}
38 changes: 38 additions & 0 deletions data/models/1x-UnResizeOnly-RCAN.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,38 @@
{
"name": "UnResizeOnly_RCAN",
"author": "kim2091",
"license": "CC-BY-NC-SA-4.0",
"tags": [
"anti-aliasing",
"restoration"
],
"description": "Purpose: Artifact Removal\n\nA version of UnResize trained on RCAN, which is faster and provides better quality than ESRGAN\n\nThis model does **not remove compression or perform deblurring**, unlike the original UnResize models. __It **only** removes scaling artifacts.__\n\nI've attached the script I used to create the dataset (it utilizes imagemagick) and the config for [traiNNer-redux](https://github.com/the-database/traiNNer-redux)",
"date": "2025-01-06",
"architecture": "rcan",
"size": null,
"scale": 1,
"inputChannels": 3,
"outputChannels": 3,
"resources": [
{
"platform": "pytorch",
"type": "safetensors",
"size": 30757598,
"sha256": "b88289e770207e634181b595845e0d240cc397714aefe659bcf9b70478b64373",
"urls": [
"https://github.com/Kim2091/Kim2091-Models/releases/download/1x-UnResizeOnly_RCAN/1x-UnResizeOnly_RCAN.pth"
]
}
],
"trainingIterations": 95000,
"trainingBatchSize": 8,
"dataset": "UltraSharpV2_Ethical, DigitalArtV3, ModernAnimation1080_v3, Kim2091's 8k Dataset V2",
"datasetSize": 13000,
"images": [
{
"type": "paired",
"LR": "https://imgsli.com/i/1cdd6669-3798-4359-b62d-0a18a38a810b.jpg",
"SR": "https://imgsli.com/i/aaa71ef9-e7c7-4af1-8aee-d653b903bd6f.jpg"
}
]
}
45 changes: 45 additions & 0 deletions data/models/2x-AnimeSharpV3-RCAN.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,45 @@
{
"name": "AnimeSharpV3_RCAN",
"author": "kim2091",
"license": "CC-BY-NC-SA-4.0",
"tags": [
"anime",
"cartoon"
],
"description": "Purpose: Anime\n\nThis release contains an ESRGAN and an RCAN version. Both provide superior quality compared to AnimeSharpV2 in nearly every scenario. It has most of the advantages of the old V2 Sharp models, while not having issues with depth of field. \n\nThe RCAN model outperforms the ESRGAN model by a significant margin, with much more consistent generation and overall better detail retention. Currently it is NOT compatible with chaiNNer, but will be available on the nightly build soon (hopefully).\n\nRCAN vs ESRGAN: https://slow.pics/c/Zqgl62Ni\n\n__Comparisons:__ <https://slow.pics/c/A2BRSa0U>\n\n\n![1729798851 305732](https://github.com/user-attachments/assets/4d7aaf33-5a39-4f75-8a05-90b90a693e49)",
"date": "2024-10-03",
"architecture": "rcan",
"size": null,
"scale": 2,
"inputChannels": 3,
"outputChannels": 3,
"resources": [
{
"platform": "pytorch",
"type": "safetensors",
"size": 31053198,
"sha256": "9c802d4d40238605d4ae8902f1f170c729ecdde142078838329b490d796292ee",
"urls": [
"https://github.com/Kim2091/Kim2091-Models/releases/download/2x-AnimeSharpV3/2x-AnimeSharpV3_RCAN.safetensors"
]
}
],
"trainingIterations": 140000,
"trainingBatchSize": 8,
"trainingOTF": false,
"dataset": "ModernAnimation1080_v3",
"datasetSize": 3000,
"pretrainedModelG": "4x-ESRGAN",
"images": [
{
"type": "paired",
"LR": "https://i.slow.pics/YAyhw81T.webp",
"SR": "https://i.slow.pics/j8BtouUm.webp"
},
{
"type": "paired",
"LR": "https://i.slow.pics/QFNEusLp.webp",
"SR": "https://i.slow.pics/tLL6gO5w.webp"
}
]
}
56 changes: 56 additions & 0 deletions data/models/2x-AnimeSharpV3.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,56 @@
{
"name": "AnimeSharpV3",
"author": "kim2091",
"license": "CC-BY-NC-SA-4.0",
"tags": [
"anime",
"cartoon"
],
"description": "Purpose: Anime\n\nThis release contains an ESRGAN and an RCAN version. Both provide superior quality compared to AnimeSharpV2 in nearly every scenario. It has most of the advantages of the old V2 Sharp models, while not having issues with depth of field. \n\nThe RCAN model outperforms the ESRGAN model by a significant margin, with much more consistent generation and overall better detail retention. Currently it is NOT compatible with chaiNNer, but will be available on the nightly build soon (hopefully).\n\nRCAN vs ESRGAN: https://slow.pics/c/Zqgl62Ni\n\n__Comparisons:__ <https://slow.pics/c/A2BRSa0U>\n\n\n![1729798851 305732](https://github.com/user-attachments/assets/4d7aaf33-5a39-4f75-8a05-90b90a693e49)",
"date": "2024-10-03",
"architecture": "esrgan",
"size": [
"64nf",
"23nb"
],
"scale": 2,
"inputChannels": 3,
"outputChannels": 3,
"resources": [
{
"platform": "pytorch",
"type": "pth",
"size": 67104146,
"sha256": "d5722d738002c1353ce4bbc3fb44fe2ecf01606c713cdb5853f772b08af84f53",
"urls": [
"https://github.com/Kim2091/Kim2091-Models/releases/download/2x-AnimeSharpV3/2x-AnimeSharpV3.pth"
]
},
{
"platform": "onnx",
"type": "onnx",
"size": 33619368,
"sha256": "fe4cbe50bfc8b20dfcb16b0935ef4dbdb64547224bee17ec2f496385bc37a71e",
"urls": [
"https://github.com/Kim2091/Kim2091-Models/releases/download/2x-AnimeSharpV3/2x-AnimeSharpV3-fp16.onnx"
]
}
],
"trainingIterations": 140000,
"trainingBatchSize": 8,
"dataset": "ModernAnimation1080_v3",
"datasetSize": 3000,
"pretrainedModelG": "4x-ESRGAN",
"images": [
{
"type": "paired",
"LR": "https://i.slow.pics/TwfO01xU.webp",
"SR": "https://i.slow.pics/4LuILsUF.webp"
},
{
"type": "paired",
"LR": "https://i.slow.pics/PsraU7Kd.webp",
"SR": "https://i.slow.pics/qdXwDmJ6.webp"
}
]
}
Loading