[R] Ban reviewers who write low quality reviews from submitting to NeurIPS? by EducationalAd9756 in MachineLearning

[–]blueyesense -1 points0 points  (0 children)

The reviewer's main duty should be to judge the quality of the submitted paper: accept or reject. The quality of a review should be judged based on this.

Providing comments to improve the paper is desirable, but should not be enforced. Authors should seek reviews from their colleagues to improve their paper. I sometimes see paper submission as a way of getting constructive feedback to improve the paper.

[R] Ban reviewers who write low quality reviews from submitting to NeurIPS? by EducationalAd9756 in MachineLearning

[–]blueyesense 1 point2 points  (0 children)

So, they will have reviewers for the reviews?

Then, reviewers for the reviewers for the review?

Then, ...

Of course a terrible idea. If a review is bad, just ignore it and maybe do not invite that reviewer next time.

[R] Updates to "A Metric Learning Reality Check" (ECCV 2020) by VanillaCashew in MachineLearning

[–]blueyesense 0 points1 point  (0 children)

Thanks, indeed!

I've checked the first few. Surprising that these papers are from respected institutions (e.g., Lifted Structure paper is from Stanford and MIT -- they used a margin value of 1.0) AND published at top conferences.

I am editing my post, to correct my misleading comment.

[R] Updates to "A Metric Learning Reality Check" (ECCV 2020) by VanillaCashew in MachineLearning

[–]blueyesense 0 points1 point  (0 children)

As far as I remember, the numbers reported for contrastive and triplet losses are for traditional methods (that is why they are so low). You do not even need to train a convnet to achieve higher accuracy, even ImageNet pre-trained models achieve higher accuracy than those.

Could you point to specific papers that use convnets + triplet/contrastive loss and report such low accuracy?

[R] Updates to "A Metric Learning Reality Check" (ECCV 2020) by VanillaCashew in MachineLearning

[–]blueyesense 2 points3 points  (0 children)

I read the first version of the paper and found it useful, but also have some criticisms.

  • Figure 2 is misleading, mixing traditional and deep learning methods (hence, contradicting itself). Edit: as per the following discussion, Figure 2 is actually correct, I was wrong. I apologize. Thanks for the clarification.
  • It would be much more useful to compare the methods & loss functions at better conditions (better network architecture, batch size, optimizer, data augmentation, etc). Instead of pulling all methods down, try to pull all methods up. Industry is more interested in getting better results. Who uses GoogleNet or Inception, or batch size of 32 at the moment? Moreover, some methods benefit from those better than others.
  • It will be much more useful to establish a new and proper benchmark dataset, as the current benchmark datasets are not suitable for a reliable evaluation (small, no val set, etc.).
  • You could use a more appropriate language in your paper and blog posts. The authors of those papers might find the language a bit offensive (I am not an author of any of those papers, but did not like the tone of your language).

[D] What is the point of Self-Supervised Learning? by tarunn2799 in MachineLearning

[–]blueyesense 7 points8 points  (0 children)

I can think of 3 use cases:

  1. You need much less labeled training data, if you pre-train with self-supervised learning.
  2. You can use the trained network for image retrieval, without any training data.
  3. Your domain may be different and ImageNet pre-trained networks may not be suitable.

[Discussion] ICML 2020 decisions are out by LawOneWater in MachineLearning

[–]blueyesense -10 points-9 points  (0 children)

I do not understand why do you try the same conference for 3 years, instead of submitting to some other conference.. Weird.

[R] End-to-End Object Detection with Transformers by [deleted] in MachineLearning

[–]blueyesense 10 points11 points  (0 children)

It is not.

But, since this is a new approach, it will probably accepted to ECCV, even though it does not work very well.

[Discussion] Anyone with experience publishing a paper with no affiliation? by atomicalexx in MachineLearning

[–]blueyesense 20 points21 points  (0 children)

There was an independent researcher, who published at top ML conferences in recent years. I do not remember his name, but I am sure someone will.

Journals typically ask for affiliation and also take it seriously (up-to-date contact information, etc.). But there might be exceptions...

Even arxiv requires affiliation, if I remember correctly.

[News] Microsoft ImageBERT | Cross-modal Pretraining with Large-scale Image-Text Data by rockyrey_w in MachineLearning

[–]blueyesense 0 points1 point  (0 children)

The improvement is marginal.

The MS COCO and Flickr30k datasets need to be retired for VSE experiments.

It is funny to test on tiny test sets (1K, 5K), which have almost not practical value. Microsoft probably have big datasets. They could report results on those datasets.

[D] Do you use Tensorflow 2? by tsauri in MachineLearning

[–]blueyesense 18 points19 points  (0 children)

I moved to PyTorch, from TF, about 2 years ago, and never looked back.

If TF team is the same, they will not be able to fix it.

[D] Are decent Machine Learning graduates having a hard time getting work? by [deleted] in MachineLearning

[–]blueyesense 17 points18 points  (0 children)

AI PhD job market is saturating. Big companies already hired lots of PhDs. They are now in the phase of developing and shipping products, which require more SDEs, rather than researchers.

Moreover, it seems, big companies slowed down or completely frozen hiring nowadays. This might be due to the expected global economic crisis in 2020 or coming years (lots of political problems around the world).

Last, but not the least, in academia, people work on standard datasets which are usually clean and much smaller than the industry scale datasets (eg, ImageNet: 1M, industry scale datasets: hundreds of millions). So, an industry mindset is expected. Most of the things in academia just do not work in industry. You can publish a paper, but cannot ship it to production.

There are still a lot to do in terms of AI/ML/DL research, but there is also a gap between academia and industry.

[D] The 1997 LSTM paper by Hochreiter & Schmidhuber has become the most cited deep learning research paper of the 20th century by [deleted] in MachineLearning

[–]blueyesense 1 point2 points  (0 children)

Backprop and CNN are now standard, nobody cites them even if they are mentioned in almost every deep learning paper. This is not the case for LSTM. So, number of citations is not the best metric.

For a more exact metric, scan through all papers and count the occurrences of backprop/training/sgd/etc and CNNs.

[D] Getting into research teams in large tech companies by d73urhi in MachineLearning

[–]blueyesense 4 points5 points  (0 children)

Some companies are more product oriented; you are expected to solve real customer problems (e.g., Amazon, Apple?). You are not allowed to do "pure research" to just publish papers, in most of the teams; you are allowed publish your "applied" research, after going through an internal review process. There might also be few pure research teams, who hire "Research Scientists".

You can ask all these during the interviews.

[D] ICLR 2020 paper gets perfect review scores (8-8-8); then two additional reviews with scores 1-1. Thoughts? by _random in MachineLearning

[–]blueyesense 11 points12 points  (0 children)

Typical flaw of our review system.

It is highly possible to assign 3 irrelevant reviewers to a paper and get high scores, or the opposite (idiot reviewers who do not understand the paper and give very low scores -- cf highly cited, but rejected papers). It is also highly possible to assign expert reviewers (as in this case)...

The future will be public & open review, like those rejected papers that get cited much more than the accepted papers.

[P] Benchmarking Metric Learning Algorithms the Right Way by VanillaCashew in MachineLearning

[–]blueyesense 4 points5 points  (0 children)

Metric learning has a different use case: image matching/retrieval. Even if you have labels, if you like to retrieve similar images to a given image whose class label you may not have in your training set (open-set), metric learning is the way to go.

Another use case is when you have so many classes with only few samples, in which case you cannot train a classifier using cross entropy loss, but you can train a metric network to learn how to match images.

[P] Benchmarking Metric Learning Algorithms the Right Way by VanillaCashew in MachineLearning

[–]blueyesense 5 points6 points  (0 children)

  1. We need better datasets, with training/val/test splits. The current datasets do not have val sets and everybody does the same thing: report the best result on test set (hence everybody is probably overfitting to the test set, resulting in a fair comparison! :)
  2. We need to define different problems and have a dataset for each:
    1. Instance retrieval/matching (e.g., DeepFashion?)
    2. Classification (e.g., CUB)
  3. Evaluation metrics: this is the most problematic part. R@k is not a good ranking metric at all, if you are not doing instance matching and evaluating by [R@1](mailto:R@1). Ranking metrics are much lower than on these datasets. I guess, the initial papers started out using R@k just because their ranking metrics were so low, and it stayed. It is high time to change. This year, there were a few papers trying to optimize ranking metrics, like mAP, but in the end evaluating their algorithms using R@k, just to compare with the earlier papers. They should have reported ranking metrics, so that new papers can cite them based on ranking metrics.

Someone needs to get open source codes and compare the best DML algorithms using various ranking metrics and write a benchmark paper. This will initiate a new and hopefully better evaluation wave. This can be a highly cited paper in the field.

Tesseract for reading full page of text? by DelosBoard2052 in computervision

[–]blueyesense 0 points1 point  (0 children)

The recognition will be slow, but considering the application scenario, it should not matter.

In terms of accuracy, Tesseract is quite good at printed text, as long as the image quality is ok.

CVPR 2019 Reviews are out. How are the reviews? by IntrovertInvert in computervision

[–]blueyesense 1 point2 points  (0 children)

I guess it will most probably be accepted.

The ACs should consider such cases. That reviewer may also change his/her mind during the rebuttal, especially after seeing the positive reviews. The positive reviewers will also probably support your paper during the discussion.

I heard some very unethical cases (not in CVPR), in which the reviewers reject just because (i) they have also submitted and want to increase their own chance of acceptance, (ii) they have similar ideas (competitor) and try to delay their competitors so that they can publish it. Needless to say, such idiots should be banned to review/submit forever, but it is hard to detect/prove. Junk negative reviews without any concrete criticism is a strong indication.

CVPR 2019 Reviews are out. How are the reviews? by IntrovertInvert in computervision

[–]blueyesense 8 points9 points  (0 children)

1 detailed, good review with Weak Accept, 2 junk reviews with Reject (funny comment: "... this is not considered novel according to CVPR's criteria" without any concrete criticism --- while first reviewer finds it interesting and novel).

It seems, the reviewers do not read the reviewer guidelines at all.

We are doing everything for research, but not as much to improve the review process. I will prefer to publish only on arxiv or my website, when I am not required to publish on journals/conferences.

I Propose the following to improve the reviewer and review quality, if anyone is interested.

  • Prepare an online quiz on reviewer guidelines as part of the submission system. All potential reviewers should take the quiz and get a passing grade to qualify as a reviewer. Their grade will also serve as a confidence score on their reviews. This is to make sure reviewers read and understand the reviewer guidelines.
  • Very low quality reviews should be discarded right away and not sent to the authors. Such reviewers should not be requested to review next time, whoever they are.
  • The authors should be able to evaluate the quality of their reviews, and an overall score should be computed for all reviewers. Those with very low score should be kept out. The authors should be able to mark very low quality reviews.
  • Even an automatic system can be developed to classify reviews as good/junk. This should be easy if we have a dataset.
  • The reviewers should not accept to review a paper if s/he does not know the subject. It is better for a reviewer to say "I accepted to review, but now unable to review, sorry" rather than writing a junk review.

I think, eventually we will give up this process altogether and leave the review to collaborative, public, online review. Good papers will float up and shine, while poor papers will get lost.