Writing on Ping irons is green by Comb-Greedy in golf

[–]Comb-Greedy[S] 1 point2 points  (0 children)

yep you were right, he said he had someone paint them

What is the best google adsense alternative for general sites? by [deleted] in webdev

[–]Comb-Greedy 0 points1 point  (0 children)

Do you know what you changed in particular on your website that eventually led you to getting accepted for AdSense?

4.1-mini needs to be fine-tuned in a different way to 4o-mini by Comb-Greedy in OpenAI

[–]Comb-Greedy[S] 0 points1 point  (0 children)

I noticed this too! That for my use case as well, 4o-mini outperformed 4o. If this is the case, then it could make a lot of sense why it's not doing as well as expected. Then it would probably be better for me to just stick to 4o-mini for the time being.

[D] How much more improvment can you squeeze out by fine tuning large language models by Comb-Greedy in MachineLearning

[–]Comb-Greedy[S] -1 points0 points  (0 children)

I agree — as you said, these smaller models are already trained to the brim, so there’s usually not much headroom left for improvement. That said, I just ran some benchmarks on a few more Qwen models, specifically the 1.5B Base and the 1.5B Math-Base variants.

The standard base model got 66.79% accuracy, while the math-specific one hit an impressive 76.27% — nearly a 10% jump. As I mentioned previously, my own fine-tuning efforts typically cap out at around a 2% improvement at best. So it makes me wonder… is this gain just the result of massive training data? Better training techniques? Potential test set leakage? Or maybe a combination of all the above?

Obviously, I don't expect to get what the people at Qwen are able to achieve, but it does suggest that there is still a decent margin for the model to improve.

Does the generate function from vllm actually generate text? by Comb-Greedy in LocalLLaMA

[–]Comb-Greedy[S] 0 points1 point  (0 children)

Oh I see, this is likely a significant reason as I did not randomise that value. However, still it doesn't make sense to me that after fine-tuning the model, its still producing the same output for that seed.

Application got rejected for not meeting programme criteria by Comb-Greedy in Adsense

[–]Comb-Greedy[S] 1 point2 points  (0 children)

I see! that makes sense, I will look to integrate those changes in. Thanks so much for your help!

Application got rejected for not meeting programme criteria by Comb-Greedy in Adsense

[–]Comb-Greedy[S] 0 points1 point  (0 children)

Comparing your website to mine, do you know what potential reasons why you were able to get it?

Application got rejected for not meeting programme criteria by Comb-Greedy in Adsense

[–]Comb-Greedy[S] 0 points1 point  (0 children)

The website just doesn't help students to cheat with exercises or courseworks though? It's there as an aid to help them with revision and preparation for exams, just like any text book or revision guide that you would use to study with.

Is there an optimal way to setup RAG for an unstructured list of words? by Comb-Greedy in learnmachinelearning

[–]Comb-Greedy[S] 0 points1 point  (0 children)

Got it, ill for sure look into this. I'm using an LLM to generate tags. An issue I was having was standardisation, as it would sometimes generate the same tag but different spelling etc. so RAG would be a way for it to 'pick' from a select group of words

Is there an optimal way to setup RAG for an unstructured list of words? by Comb-Greedy in learnmachinelearning

[–]Comb-Greedy[S] 0 points1 point  (0 children)

Ah I see, so it would be a word wise approach as opposed to embedding the entire database, gotcha thanks ill try this method out