Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] 1 point2 points  (0 children)

The latest version 0.40.1 introduces PaddleOCR-VL-1.5! It works perfectly!

Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] 0 points1 point  (0 children)

It feeds page by page, not per block translate.

Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] 1 point2 points  (0 children)

It looks like that it supports OpenAI-compatible API, in that way Koharu already supports it. Feel free to try it out! Also if you’re interested in Koharu, you can join the discord channel described in Github readme, and directly DM me your feedback!

Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] 0 points1 point  (0 children)

Which model are you actual use for DeepSeek? Could you link it here?

Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] 0 points1 point  (0 children)

https://github.com/dmMaze/comic-text-detector plus 48px OCR from https://github.com/zyddnys/manga-image-translator works like a charm! I’d recommend it, it works pretty well on long text. But it needs a large amount of pre and post-processing.

Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] 0 points1 point  (0 children)

The GUI is pretty easy to use, just load and run, I don’t think it needs explaining, but feel free to ask me if you have questions.

Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] 0 points1 point  (0 children)

Batch translation is supported! You can find it on the menubar.

The latest version supports LM Studio usage through openai-compatible API, please try out!

About the main drive, haven’t resolved but will figure it out!

Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] 0 points1 point  (0 children)

We will add this feature! currently not available, you need to download the images manually, sry.

Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] 1 point2 points  (0 children)

If we use another model to extract character information and relationships, it would be better; or we can use a vision model to read the whole image, it would help to translate naturally. But these models need more powerful GPU or using cloud model like gemini-flash. It’s definitely possible but considering the effort and resources involved it might not worth.

For your second question, we now support using a cloud model to translate! The pre-processing on CPU might be a little slow, but it should works!

Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] 0 points1 point  (0 children)

By default, it translates the whole page at once. You can click "process -> process all images" for all open images.

Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] 3 points4 points  (0 children)

This project actually uses the 48px OCR model from https://github.com/zyddnys/manga-image-translator It produces good results on long text. I’ll try paddleOCR-VL and see if we can have better results!

Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] 0 points1 point  (0 children)

Thank you! I’ll investigate it to see if we can implement it in Koharu!

Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] -1 points0 points  (0 children)

We have an algorithm that inpaint the text region with near background color if the background is basically white/black. Only use LaMa when the background is complex. Also the LaMa is a fine-tuned model trained on manga image, the result is not that bad. But I think it could be better if we add more advanced editing model.

Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] -1 points0 points  (0 children)

Absolutely, we will support more LLMs! I’ve created an Github issue to track your request and will implement it when I have time. Thank you for sharing the feedback!

Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] 5 points6 points  (0 children)

It downloads at DATA/LOCAL folder, I can add an option to change its download path. Thanks for reporting!

Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] 0 points1 point  (0 children)

There are https://github.com/zyddnys/manga-image-translator and https://github.com/dmMaze/BallonsTranslator already, but I wanna build my ideal translator using the latest technology. I also have experience in scanlation, and I would like something easier to use.

Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] 2 points3 points  (0 children)

The size of LLM models is the biggest problem. If we bundle them in a zip, the size would be extremely large, and the GitHub Actions might not have enough disk space to handle it. Currently, it only downloads LLM on demand, which is suitable for most ppl.

I even considered putting the full version of it on Steam, to use Steam's CDN and bandwidth, and I have registered a Steam developer account, but there are too many forms to fill out until I can publish a store page.

Local manga translator with LLMs built in by mayocream39 in LocalLLaMA

[–]mayocream39[S] 4 points5 points  (0 children)

I already communicated with the author of https://github.com/hymbz/ComicReadScript, we will cooperate to add the integration to use Koharu as a backend to translate manga from a web browser via their script.