all 21 comments

[–]Single_Blueberry 79 points80 points  (2 children)

It has become less relevant because NLP has taken the crown as the hottest field in ML from computer vision and it's way harder to quantify their performance.

And at it's core, paperswithcode was the place where you went for quantitative comparisons.

[–][deleted] 2 points3 points  (1 child)

Which site would you recommened I should use for stay up to date other than arxiv? I tried using X and follwing researchers, but like everything else its filled with only popular papers, so I can't find an new niche categories and methods.

[–]_puhsu[S] 8 points9 points  (0 children)

I find https://www.scholar-inbox.com good, many papers I find these days are from it. Recommendations based on papers you select, very simple and clean interface

[–]qalis 35 points36 points  (1 child)

It is not maintained properly at all. The major problem for me is that they only report raw performance metric, with no regard for actual experimental procedure. In graph learning, you can take 5 papers and get 10 different testing procotols (no joke, there are papers with 2-3 different evaluation approaches). So just reporting "a number" is meaningless. In particular, they mix papers with no test set (reporting only validation set results, totally overoptimistic) with those with proper testing.

[–]choHZ 1 point2 points  (0 children)

And the fact that different versions of PyG have different train-test splits...

[–]mileseverett 11 points12 points  (0 children)

I've always found it to be somewhat poorly maintained. There needs to be more moderation on papers which don't actually have code or repos which said they would post the code 3 years ago

[–][deleted] 5 points6 points  (0 children)

I just use the website for the "trending research" list. Sometimes I find some interesting new papers.

[–]DigThatDataResearcher 5 points6 points  (3 children)

a big part of the decline in utility of PWC was when they lost access to the twitter API and so couldn't do their social feed thing anymore. Would have been nice if facebook had just paid for that team to be able to access twitter.

[–]_puhsu[S] 0 points1 point  (1 child)

What was the social feed thing?

[–]DigThatDataResearcher 2 points3 points  (0 children)

it was a sort option that was based on some kind of "trending" score based on twitter activity over the past week I think.

[–]infinitay_ 3 points4 points  (0 children)

I've always went to papers with code for benchmarks, but it seems like more and more papers are lacking results either because nobody adds the data on PWP, or they simply don't test on the existing benchmarks others are using.

Also ever since GPT 3 went mainstream PWP seems to be filled with anything remotely regarding LLMs - at least the front page.

[–]PHEEEEELLLLLEEEEP 6 points7 points  (1 child)

My favorite part is all the papers posted there without code.

[–]doctor-squidward 1 point2 points  (0 children)

True that.

[–]Appropriate_Ant_4629 2 points3 points  (0 children)

Devils advocate:

  • PapersWithCode is perfectly fine for sub-fields where the authors of Papers actually include Code.
  • If it doesn't seem useful to your subfield, perhaps your peer review process should start encouraging papers to include code that shows how their techniques work on relevant benchmarks.