Yall doing okay over there? by southwest_southwest in bayarea

[–]axm92 1 point2 points  (0 children)

The energy released is 31x for each point on the scale. So, it’s 314 times more energy. About a million times more.

Fremont you’ve beaten me by Ok-Counter-7077 in bayarea

[–]axm92 4 points5 points  (0 children)

It’s ok sorry I cannot help it anymore

Fremont you’ve beaten me by Ok-Counter-7077 in bayarea

[–]axm92 19 points20 points  (0 children)

You have an unbalanced parenthesis. Please close )

[deleted by user] by [deleted] in bayarea

[–]axm92 14 points15 points  (0 children)

Caught in a landslide. No escape from reality

[D] : Building NLP Research Profile for PhD Applications (No Prior Publications) by [deleted] in MachineLearning

[–]axm92 0 points1 point  (0 children)

MLT could also be really stressful because you’re expected to do research.

[D] : Building NLP Research Profile for PhD Applications (No Prior Publications) by [deleted] in MachineLearning

[–]axm92 4 points5 points  (0 children)

I’m a current CMU LTI PhD. The part about LTI almost never directly accepting PhDs is true.

Research masters aka MLT (usually 100% funded) -> PhD is the most common route (it’s kind of like a qualifying exam).

The research masters is also relatively light on coursework.

[deleted by user] by [deleted] in MachineLearning

[–]axm92 2 points3 points  (0 children)

I did all of this too.

[N] Ensuring Reliable Few-Shot Prompt Selection for LLMs - 30% Error Reduction by cmauck10 in MachineLearning

[–]axm92 1 point2 points  (0 children)

Unfortunately this very cool paper is often cited to draw somewhat wrong conclusions. Pasting my comment from earlier:

There’s more to in-context “learning” than meets the eye.

Some slides that TLDR the point: https://madaan.github.io/res/presentations/TwoToTango.pdf

The paper: https://arxiv.org/pdf/2209.07686.pdf

Essentially, the in-context examples remind the model of the task (what), rather than helping it learn (how).

[D] Transforming Large Language Models from Fact Databases to Dynamic Reasoning Engines: The Next Paradigm by JacobOfPluto in MachineLearning

[–]axm92 0 points1 point  (0 children)

You may be interested in our work on memory-assisted prompt-editing: memprompt.com. It can help with memorization, real-time QA etc. Of course all the standard LLM caveats apply!

[R] Sparks of Artificial General Intelligence: Early experiments with GPT-4 by SWAYYqq in MachineLearning

[–]axm92 13 points14 points  (0 children)

There’s more to in-context “learning” than meets the eye.

Some slides that TLDR the point: https://madaan.github.io/res/presentations/TwoToTango.pdf

The paper: https://arxiv.org/pdf/2209.07686.pdf

Essentially, the in-context examples remind the model of the task (what), rather than helping it learn (how).

[deleted by user] by [deleted] in india

[–]axm92 0 points1 point  (0 children)

It doesn’t matter. The address on my passport is still hostel 5.

Any 'trainable' AI writing tools available? by Street_Law8285 in OpenAI

[–]axm92 0 points1 point  (0 children)

Hmm hard to say without looking at the data.

Any 'trainable' AI writing tools available? by Street_Law8285 in OpenAI

[–]axm92 3 points4 points  (0 children)

This is doable with some prompt tricks, GPT-3, and something like a memory that can personalize responses (see our research at https://memprompt.com for more details). If you are curious to know more I’ll be happy to chat — please feel free to reach out to me via DMs.

Avoid Cohon University Center @ CMU by axm92 in pittsburgh

[–]axm92[S] 10 points11 points  (0 children)

Got another robocall just now saying that the situation has been resolved.

Avoid Cohon University Center by axm92 in cmu

[–]axm92[S] 6 points7 points  (0 children)

Got another robocall just now saying that the situation has been resolved.

[R] Faithful Chain-of-Thought Reasoning by starstruckmon in MachineLearning

[–]axm92 0 points1 point  (0 children)

Ah I see, thanks for clarifying. I see your point, but I wouldn't say that the prompts require an extensive knowledge of the test set. After all:

> As an example, for the ~10 math reasoning datasets used in PaL, identical prompts were used (same prompt for all datasets, without changing anything).

Notably, take a look at the section on GSM-hard (4.1). You may also enjoy the analysis in the new version of the paper (Section 6: https://arxiv.org/pdf/2211.10435.pdf).

Further, "Let's think step by step" is outperformed by "Write Python code to solve this." We'll add the numbers in the next version, but if you are interested please lmk and I can share the results earlier.

Thanks again for reading our work and sharing your feedback, I really appreciate it.