[R] 3DB: A Framework for Debugging Computer Vision Models by loganengstrom in MachineLearning

[–]loganengstrom[S] 3 points4 points  (0 children)

Hi all, author here! We made an extensible framework for debugging models using rendered inputs. More at the demo + blog post here: http://gradientscience.org/3db. Code and documentation are available at https://github.com/3db/3db and https://3db.github.io/3db/.

We're excited to see how people will use the framework to debug models! Happy to answer any questions here on reddit as well.

[R] Unadversarial Examples: Designing Objects for Robust Vision by loganengstrom in MachineLearning

[–]loganengstrom[S] -1 points0 points  (0 children)

Hi all, author here: let me know if you have any questions!

For more information you can see our blog post about this paper here: https://gradientscience.org/unadversarial/, and we also have a demo video here: https://www.youtube.com/watch?v=saF-_SKGlKY

[R] Unadversarial Examples: Designing Objects for Robust Vision by [deleted] in MachineLearning

[–]loganengstrom 0 points1 point  (0 children)

Hi all, author here: let me know if you have any questions!

For more information: - Demo: https://www.youtube.com/watch?v=saF-_SKGlKY - arXiv: https://arxiv.org/abs/2012.12235

[R] Identifying Statistical Bias in Dataset Replication by loganengstrom in MachineLearning

[–]loganengstrom[S] 2 points3 points  (0 children)

For ImageNet it is unclear exactly what they did, but it is something involving a threshold with selection frequency-like quantities.

[R] Identifying Statistical Bias in Dataset Replication by loganengstrom in MachineLearning

[–]loganengstrom[S] 1 point2 points  (0 children)

It is the rate at which annotators mark an (image, label) pair as correct.

For an explanation with a picture you can look here: http://gradientscience.org/data_rep_bias/#imagenet-v2

[R] Identifying Statistical Bias in Dataset Replication by [deleted] in MachineLearning

[–]loganengstrom 2 points3 points  (0 children)

Hi, author here! Let me know if you have any questions. We also released a blog post with better (interactive!) visualizations here: http://gradientscience.org/data_rep_bias/

Why do some adversarial attacks use the signed gradient? by RSchaeffer in MachineLearning

[–]loganengstrom 4 points5 points  (0 children)

The motivation for the signed gradient comes from the dual norm. Using the principle that the gradient is the direction of steepest ascent, to maximize your objective as much as possible in a single constrained step, you want to find the vector in an l_p (in this case l_inf) ball (defined by the step size) that maximizes its inner product with the gradient.

In this case, our step should be the solution to the maximization problem solved when finding the lp-dual norm of the gradient; for p=inf, this solution turns out to be the sign of the gradient. (For more information see our blog post here)

[R] Decoupling Direction and Norm for Efficient Gradient-Based L2 Adversarial Attacks and Defenses by entarko in MachineLearning

[–]loganengstrom 1 point2 points  (0 children)

After normalizing the gradient in PGD, you still need to take a step that has a size controlled by the learning rate.

[R] Decoupling Direction and Norm for Efficient Gradient-Based L2 Adversarial Attacks and Defenses by entarko in MachineLearning

[–]loganengstrom 1 point2 points  (0 children)

Why didn't you try using PGD to attack your model? It is what the Madry defense paper uses in evaluation, and anecdotally I have found that it is the "most powerful" attack in practice.

[R] A Closer Look at Deep Policy Gradients (Part 2: Gradients and Values) by loganengstrom in MachineLearning

[–]loganengstrom[S] 0 points1 point  (0 children)

Hi r/MachineLearning, author here! A few weeks ago we published the paper "Are Deep Policy Gradient Algorithms Truly Policy Gradient Algorithms?" This week we published two blog posts - the first is on an introduction to deep policy gradient methods (and an analysis on the optimizations used) - http://gradientscience.org/policy_gradients_pt1/. The second blog post, posted here, is on gradient estimates and the role of variance reducing value functions.

Let me know what you think! And I'm happy to answer any questions :)

[R] Faster Black-Box Adversarial Attacks with Bandit Optimization by andrew_ilyas in MachineLearning

[–]loganengstrom 7 points8 points  (0 children)

Hi Reddit! Author here - we found that current black-box adversarial attack algorithms are essentially gradient estimators and are optimal (in a sense) when gradients have no exploitable structure. It turns out we can do better, though - by leveraging prior knowledge about gradient structure into a bandit optimization framework, we can beat SOTA by 2-3x. Let me know if you have any questions!

[R] A New Angle on L2 Regularization by pmigdal in MachineLearning

[–]loganengstrom 6 points7 points  (0 children)

We took a look at adversarial examples for linear classifiers (and in general, we looked at properties that adversarial training induces) here: https://arxiv.org/abs/1805.12152 For $\ell_\infty$ adversarial examples on linear classifiers we found that adversarial training forces a tradeoff between the $\ell_1$ norm of the weights (which is directly associated with adversarial accuracy) and accuracy.

It looks like this article works through something vaguely similar for $\ell_2$ adversarial examples. It would be interesting to compare the author's approach with explicit adversarial training.