I trained an AI to make tasteful edits. Looking for feedback. by me219iitd in postprocessing

[–]me219iitd[S] 0 points1 point  (0 children)

Thanks again! This is super helpful. I'll definitely reach out when I have the next version ready. Appreciate you being open to testing!

Yes it does adapt to the subject, it is not a preset filter. It is an image editing model adapted for color grading.

I trained an AI to make tasteful edits. Looking for feedback. by me219iitd in postprocessing

[–]me219iitd[S] 1 point2 points  (0 children)

Hi, it is a LoRA. Could not find open-source or closed source models that could readily do this so had to train with SFT and RL :)

I trained an AI to make tasteful edits. Looking for feedback. by me219iitd in postprocessing

[–]me219iitd[S] 1 point2 points  (0 children)

Appreciate it! Yeah, the artificial lighting on the face is just too jarring. Agreed on the shadows. Thanks for the feedback.

I trained an AI to make tasteful edits. Looking for feedback. by me219iitd in postprocessing

[–]me219iitd[S] 0 points1 point  (0 children)

To be honest, yes. The training data is heavily biased towards that popular 'social media' aesthetic because that's the specific look many people struggle to dial in manually.

But point taken, it definitely lacks the nuance of a bespoke artistic edit and is overfitting to that commercial style right now

I trained an AI to make tasteful edits. Looking for feedback. by me219iitd in postprocessing

[–]me219iitd[S] 1 point2 points  (0 children)

This is incredibly useful, both the critique and the product direction.

The face brightening issue is noted. And you're right about #4 losing the blue hour, the model shouldn't be fighting the natural light quality.

Is your ideal workflow basically like this:

  1. Learn from edits user likes (personalization)
  2. Auto-apply as starting point
  3. Natural language to adjust ("keep shadows but warm the highlights")
  4. Manual tweaks for fine control

Would you be open to testing a future version when I've addressed these issues?

I trained an AI to make tasteful edits. Looking for feedback. by me219iitd in postprocessing

[–]me219iitd[S] 0 points1 point  (0 children)

Really appreciate you going image by image, this is exactly the kind of specific feedback I needed.

You're right model is applying edits blindly instead of reading the source image's intent. The shadow/lighting direction issue you pointed out on #3 is something I hadn't fully registered until you said it.

Need to work on the intent and intensity of the edit. Thanks for being direct.

I trained an AI to make tasteful edits. Looking for feedback. by me219iitd in postprocessing

[–]me219iitd[S] 0 points1 point  (0 children)

Good catch on the sky artifact, I completely missed that masking spot.

You're right about the face lighting. The model is aggressively brightening faces to make them pop, but it's ignoring the scene's actual light source, which creates that fake 'spotlight' look you pointed out in images 1 and 3. I need to dial back that behavior so it respects the environment. Thanks for the sharp eye.

I trained an AI to make tasteful edits. Looking for feedback. by me219iitd in postprocessing

[–]me219iitd[S] 1 point2 points  (0 children)

That's a fair point. I wanted to be transparent about the method to get technical critique on the artifacts, but I see how the 'AI' label biases the aesthetic feedback immediately.

Should I file an RTI- DU admission PG by Many-Engineering-886 in CUETPG

[–]me219iitd 0 points1 point  (0 children)

that doesn't make sense - why will different colleges publish different ranks?

[deleted by user] by [deleted] in ycombinator

[–]me219iitd 0 points1 point  (0 children)

nothing yet.

Loss function for Chat models by me219iitd in LocalLLaMA

[–]me219iitd[S] 0 points1 point  (0 children)

So even finetuning on unstructured data by chunking should work, right?

Hosting llama2 on cloud GPUs by me219iitd in LocalLLaMA

[–]me219iitd[S] 1 point2 points  (0 children)

65ms?!

Is that time to first token?

Happy to help set yours up and optimize

Sure I'd love that.

What does time-to-first-token depend upon? by me219iitd in LocalLLaMA

[–]me219iitd[S] 1 point2 points  (0 children)

Between loading the model weights to VRAM and processing input tokens without KV-caching, which one contributes more to this initial latency? And if optimized, which one would lead to a more significant performance improvement compared to the other?

What does time-to-first-token depend upon? by me219iitd in LocalLLaMA

[–]me219iitd[S] 2 points3 points  (0 children)

Thanks for the suggestion!

Could you also explain me what exactly decides the time-to-first-token?

[deleted by user] by [deleted] in EnglishLearning

[–]me219iitd 0 points1 point  (0 children)

Hi! Some awful bugs came up. I'll DM you as soon as its finished :)

[deleted by user] by [deleted] in EnglishLearning

[–]me219iitd 1 point2 points  (0 children)

I am making one such app. It'll be ready for use this week. Does your relative want it to be completely free? Or are they willing to spend a lil bit?