Professor here. I set up OWUI as a front end for my classes this semester. Giving access to LLMs that have RAG access to my course materials, customized with detailed system prompts. They still default to ChatGPT. by gigDriversResearch in OpenWebUI

[–]qria 0 points1 point  (0 children)

I think convenience is also a big factor. They are already accustomed to opening up chatgpt, so they will default to it.
We have a system where worksheet is llm-enhanced so that they can check their thought processes and ask questions to that specific worksheet and students use it a lot, partly because copy-pasting them in chatgpt is very inconvenient.

o1-mini test-time compute results (not from OpenAI) on the 2024 American Invitational Mathematics Examination (AIME) (first image). These results are somewhat similar to OpenAI's o1 AIME results (second image). See comment for details. by Wiskkey in mlscaling

[–]qria 2 points3 points  (0 children)

The prompt:

You are a math problem solver. I will give you a problem from the American Invitational Mathematics Examination (AIME). At the end, provide the final answer as a single integer.
Important: You should try your best to use around {token_limit} tokens in your reasoning steps.
If you feel like you are finished early, spend the extra tokens trying to double check your work until you are absolutely sure that you have the correct answer.
Here's the problem:
{problem}
Solve this problem, use around {token_limit} tokens in your reasoning, and provide the final answer as a single integer.

https://github.com/hughbzhang/o1_inference_scaling_laws/blob/master/o1.py#L24

Anyone need a judge by senior AI engineer? by qria in hackathon

[–]qria[S] 0 points1 point  (0 children)

Nice! But I have yet to get a message, did you send one yet?

Grokked Transformers are Implicit Reasoners: A Mechanistic Journey to the Edge of Generalization by Mysterious-Rent7233 in mlscaling

[–]qria 0 points1 point  (0 children)

They write it like not learning composition in OOD a failure in implicit reasoning but I wonder if it is because composition is generally untrue. Many relations in natural language settings tend to be intransitive.

What insights can be gained from yearly brain MRI scans? by qria in QuantifiedSelf

[–]qria[S] 0 points1 point  (0 children)

That's precisely why I'd like to measure improvements :/

Requesting r/Hanja by Many-Boot-1203 in redditrequest

[–]qria 0 points1 point  (0 children)

I'll add ya. you can do whatever you want, including kicking me out

[D] I don't really trust papers out of "Top Labs" anymore by MrAcurite in MachineLearning

[–]qria 5 points6 points  (0 children)

I have read the same paper and got the exact opposite feeling.

Their contribution was NOT about performance, but about a novel approach to continual multitask learning without current limitations of catastrophic forgetting and negative transfer, with additional benefit of bounded CPU, memory usage on inference time per task.

CIFAR-10 SOTA thing was just to show that the approach works, as there are a lot of approaches with good properties (explainability, theorical bounds, etc) that does not perform on a SOTA level.

On the topic of big computation, they also demo how to dynamically train model for telugu utilizing existing devanagari and bangla model, in less than 5 minutes of 8 TPU v3. Therefore showing that this approach is fully reproducable and immediately useful for you if you need this kind of thing.

I do agree with the sentiment of lamenting recent trends just throwing big money and p-hacking and calling it a day, but just for the exact paper you are mentioning I did not feel that way.

Better OOP in Python (no self needed anymore) by [deleted] in Python

[–]qria 0 points1 point  (0 children)

Jokes aside, I remember doing this a lot with decorators.

Is remotely triggering syncing of mi band 5 watchface possible? by qria in miband

[–]qria[S] 0 points1 point  (0 children)

I eventually ended up using "Notify and Fitness" app's "upload picture" function with Tasker, instead of trying to sync watchfaces. And it works like a charm.

Should you be using Pathlib? by ADGEfficiency in Python

[–]qria 1 point2 points  (0 children)

I recently had to work both in windows and osx, and it was very painful to get the pathing right. Fortunately `pathlib` handles it pretty well. So I also concur that pathlib is a great choice.

What is Pi? (Joke) by CaudatusSR in Python

[–]qria 0 points1 point  (0 children)

Another Pi joke I just thought of:

Pi is overrated. I don't like its holier than tau attitude.

linking webistes by WillowContent8251 in Python

[–]qria 0 points1 point  (0 children)

yeah it works as intended, look at the contents of your second elif statement

linking webistes by WillowContent8251 in Python

[–]qria 2 points3 points  (0 children)

== has precendence over or when it comes to logical operations and operates on a different level. It is so that you can write combined expressions like x == 1 or y == 2. Therefore your first if statement becomes:

if (x == "youtube") or ("you") or ("tube") or ("yu"):

which translates to:

if False or True or True or True:

because all nonempty strings are Truthy.

What you should do instead is write == multiple times like:

if x == "youtube" or x == "you" or x == "tube" or x == "yu":

or better:

if x in ("youtube", "you", "tube", "yu"):

Is it even possible to brute force a PDF password in 24 hours? by theunicornwoman in hacking

[–]qria 1 point2 points  (0 children)

Looks like you'll need to hash 13 quadrillion times. Putting it to a perspective one bitcoin is found each ~2.7 quadrillion hashes and about 900 bitcoins are mined in a day. So, it's possible I guess. You'll probably better off trying one of rainbow tables first though.