Hey I don't know, but sheryl have this fan translationaction game with pixel art? +1 for sheryl
Ty and nice, I don't know if it's gonna pass, but if you suddenly have the link to download it can you send it?A translation was already completed for Sheryl. it is being reviewed for Rule 7.
I think there is a tranlsation out there for the first one too.
third game: https://f95zone.to/threads/translation-requset-sheryl-mega-download.87836/
second and third one doesn't have any translation.
Edit 2: Sheryl has now some community translation beign made now, one even beign in the test of the rule 7 so good I guess, now the only game that doesn't have any near translation it's the third one
Erm... I'm a bit confused what you mean with it doesn't have any translation.action game with pixel art? +1 for sheryl
The thread is full of weird translation attempts and if it's getting an official translation anyway I don't really want to work on it.Hi, I would like to make 3 request
first game: https://f95zone.to/threads/translation-requset-rpg-konoha-lolique2gaiden-mega-download.103577/
Edit: first game it's actually having an official translation for steam, it's official annunced by otaku plan it doesn't have any date for now, but for what I've seen it gotta be a censored version for steam, and l don't know at the same time or later, but a +18 patch
I'll look into it. Edit: Done.
Understandable that it's a work in progress. Looking through the results, it seems like it really struggled with smaller fonts, especially when complicated kanji are involved, or distinguishing between something like ぽ and ぼ. I'm guessing that's one point of difficulty, especially if it fails in a 1600xsomething game; most rpgmaker games are already tiny.Alrighty. I had to implement it replacing the alpha channel before scanning, but now it's somewhat working.
I've attached the result. Please let me know what you think.
But ofc keep in mind that this happened in like 40 seconds with offline neural networks that I'm running in my ram, so obviously you can't favorably compare it to something like Google Lens or a manual translation.
Edit: I should also mention that your example was pretty much the worst case scenario of half transparent short text. It works a lot better for bigger blocks of text.
I've tried it, but since the issues are with the OCR and not SugoiV4(The MTL) here's the actual Japanese output of each picture instead of it being drawn on the images. (The numbers are the coordinates of the corners around the detected text block.)Understandable that it's a work in progress. Looking through the results, it seems like it really struggled with smaller fonts, especially when complicated kanji are involved, or distinguishing between something like ぽ and ぼ. I'm guessing that's one point of difficulty, especially if it fails in a 1600xsomething game; most rpgmaker games are already tiny.
It also struggles with 0, some of the images from a-1_18 to 34 drop the zero.
When the ocr does successfully recognize the text, the en text returned at least somewhat catches the gist of the text (stripped of context sure, but that's something all mtls struggle with).
If you're open to trying something out, could you run the same texts again with the 4x upscaled ones below? I don't expect it to be wildly successful since the images have artifact/strokes blending together issues, but it's worth checking out I think.
The workflow I'm thinking of for this is to upscale tiny images using chainner, apply your ocr, then downscale them again with chainner.
Upscaled images:You must be registered to see the links
Yeah, I have experience with Sugoi and it's a decent mtl tool. I got that the ocr had trouble recognizing the texts, which is why I thought it might work better with the upscaled images. I have absolutely no programming experience, so improving the image input (via upscaling) was my suggestion to improve the ocr process.I've tried it, but since the issues are with the OCR and not SugoiV4(The MTL) here's the actual Japanese output of each picture instead of it being drawn on the images. (The numbers are the coordinates of the corners around the detected text block.)
If you want the actual MTL results instead just tell me, I just thought this would work better for you.
I really appreciate the help.
I'm not sure what feedback you need exactly to improve/train the ocr. What do you need?Understandable that it's a work in progress. Looking through the results, it seems like it really struggled with smaller fonts, especially when complicated kanji are involved, or distinguishing between something like ぽ and ぼ. I'm guessingthat'sthat the blurry texts are one point of difficulty, especially if it fails in a 1600xsomething game; most rpgmaker games are already tiny.
Upscaled images:You must be registered to see the links
Honestly I just wanted to know now if you think that upscaling it actually improved it now or not. (Those Json were made using your upscaled versions.)Yeah, I have experience with Sugoi and it's a decent mtl tool. I got that the ocr had trouble recognizing the texts, which is why I thought it might work better with the upscaled images. I have absolutely no programming experience, so improving the image input (via upscaling) was my suggestion to improve the ocr process.
Glancing through the json files though, I'm pretty sure what I wrote below is happening.
I'm not sure what feedback you need exactly to improve/train the ocr. What do you need?
I don't really want to work on an outdated version.https://f95zone.to/threads/translation-request-a-virgin-awakens-her-sexuality.131942/
Can you try translating that one Shisaye? There's a mega download link there but it's outdated and the one on hentai sharing is also not up to date sadly.
Aw bummer, understandable though. I'll go look for something else then.I don't really want to work on an outdated version.
According to dlsite those weren't just bugfixes since 1.02, but actual new content added.
Ohhh , that's my bad then, I'm the one who misunderstood. I thought the json files you sent were from the original images.Honestly I just wanted to know now if you think that upscaling it actually improved it now or not. (Those Json were made using your upscaled versions.)
Or if you would say the detection is pretty much the same.
Also if you see potential in general, have ideas, before I run it on all the picture based text missing from my SLRMTL projects.
I honestly thought I was pretty much done with it because my tests before that actually worked really well, while you now managed to make it look like garbage.
So now I'm trying to come up with ways to improve the OCR on those types of pictures.
To explain how much of a difference I mean, here's some examples from the game "Figure of a Hero" which I originally used to test the system.
I've applied it to the original pictures because that's what would have to happen anyway, but I used the upscaled version for the scanning. (I just divided all box coordinates by 4, it's not perfect, but kinda worked.)Ohhh , that's my bad then, I'm the one who misunderstood. I thought the json files you sent were from the original images.
Could you send over the mtl overlaid on the upscaled images? It still struggled with the first image, but it got more accurate readings for a few that were mistranslated in the original. I want to see if the final results actually got better.
Hm... I wonder if ChaiNNer does a better job than RealESRGANx4plus. That's what I used for upscaling till now.After looking through the images, here're my thoughts.
1. Upscaling - Out of all the mistranslated lines in the original ocr, about 30-40% were fixed in the upscaled ocr. It also improved some lines that were tled somewhat accurately. However, the upscaled ocr did sometimes get something wrong that the og ocr got right, but less often than the corrections it gave, and even then only for katakana stuff. The zeros were also accounted for better here, but that's probably an issue of the character used for them, and not size/blurriness.
Workflow suggestion: Upscale large images with different blocks of texts in them. Images that contain only texts (e.g. attached below) are okay to use as is, unless really blurry/low-res.
2. Deepl vs Sugoi - Deepl does better translating UI labels like variable or item labels (Affection, level, time of day, sex toy, etc.). When the OCR failed to give the proper kanji for a word, Deepl was able to correct it somehow. For sentences though, sugoi does better.
3. The overlaid text is too small on some of the upscaled images for some reason, e.g. attached below.
I used ChaiNNer to upscale the images since it was able to preserve the alpha channels in the images. It's python-based, so you could automate that too, and it has native support for batch upscaling images.
def word_wrap(image, ctx, text, roi_width, roi_height): #this likes crashing :C
"""Break long text to multiple lines, and reduce point size
until all text fits within a bounding box."""
mutable_message = text
iteration_attempts = 100
def eval_metrics(txt):
"""Quick helper function to calculate width/height of text."""
metrics = ctx.get_font_metrics(image, txt, True)
return (metrics.text_width, metrics.text_height)
while ctx.font_size > 0 and iteration_attempts:
iteration_attempts -= 1
width, height = eval_metrics(mutable_message)
if height > roi_height:
ctx.font_size -= 0.75 # Reduce pointsize
mutable_message = text # Restore original text
elif width > roi_width:
columns = len(mutable_message)
while columns > 0:
columns -= 1
mutable_message = '\n'.join(wrap(mutable_message, columns))
wrapped_width, _ = eval_metrics(mutable_message)
if wrapped_width <= roi_width:
break
if columns < 1:
ctx.font_size -= 0.75 # Reduce pointsize
mutable_message = text # Restore original text
else:
break
if iteration_attempts < 1:
raise RuntimeError("Unable to calculate word_wrap for " + text)
return mutable_message
Yeah... I'm not a programmer so I really don't get any of the lines of code there.Hm... I wonder if ChaiNNer does a better job than RealESRGANx4plus. That's what I used for upscaling till now.
As for the font size, that's a long story and most likely not actually a problem with the fontsize but with the box size because I have to try to make it always fit the box regardless of the overall scale.
I guess that box somehow got messed up and turned out way too small.
If you're curious this is how my script for determining linebreaks and font size looks like. (Taken from the Wand documentation.)
Oh neat. I'll try those out.Yeah... I'm not a programmer so I really don't get any of the lines of code there.
As for upscaling, ChaiNNer isn't an upscaling model itself, it's the GUI used to apply the upscaling models to images. I used 4x_-Ultrasharp since it was best at preserving kanji strokes out of those I tested. There're models better for specialized use out there, and Ultrasharp or Animesharp will probably be better than realesregan.
You must be registered to see the links
You must be registered to see the links
You must be registered to see the links
There's also this specialized for lowres text,You must be registered to see the links, but it wasn't any better than Ultrasharp.
You must be registered to see the links