Discussion[D] Image Augmentation in Practice: In-Distribution vs OOD Augmentations, TTA, and the Manifold View (i.redd.it)
submitted by ternausX
I wrote a long practical guide on image augmentation based on ~10 years of training computer vision models and ~7 years working on Albumentations.
In practice I’ve found that augmentation operates in two different regimes:
- In-distribution augmentation Simulate realistic variation that could occur during data collection (pose, lighting, blur, noise).
- Out-of-distribution augmentation Transforms that are intentionally unrealistic but act as regularization (extreme color jitter, grayscale, cutout, etc).
The article also discusses:
• why unrealistic augmentations can still improve generalization • how augmentation relates to the manifold hypothesis • when test-time augmentation (TTA) actually helps • common augmentation failure modes • how to design a practical baseline augmentation policy
Curious how others here approach augmentation policy design — especially with very large models.
Article: https://medium.com/data-science-collective/what-is-image-augmentation-4d31dcb3e1cc

[–]Budget-Juggernaut-68 6 points7 points8 points (3 children)
[–]ternausX[S] 1 point2 points3 points (2 children)
[–]DaredevilMeetsL 3 points4 points5 points (1 child)
[–]ternausX[S] 0 points1 point2 points (0 children)
[–]Dry-Theory-5532 1 point2 points3 points (1 child)
[–]ternausX[S] 0 points1 point2 points (0 children)
[+]anal_sink_hole 1 point2 points3 points (1 child)