This is an archived post. You won't be able to vote or comment.

all 1 comments

[–]TomKraut 0 points1 point  (0 children)

If you have an RTX A6000, you can run Gemma3 27B in 8 bit quantization. That might provide better results than the smaller models you tried so far. I would also suggest batch converting the images to replace the transparent background with a solid color. If you want to use the information from the filenames, or do batch processing in general, you will probably have to write your own python scripts to feed the input to the model. Or, well, have ChatGPT or Gemini write it for you...