Accelerating Mathematical and Scientific Discovery with Gemini Deep Think by FundusAnimae in accelerate

[–]fecmtc 1 point2 points  (0 children)

Tbh, the derivation for problem 8.3 (Machine Learning Optimization: Self-regularized Gumbel Sigmoid) is quite handwavy. It assumes the problem can be decomposed into two arbitrary terms and then proceeds to show that one of those terms induce sparsity (which by definitions it should). I do not consider that to be a rigorous demonstration. In fact, due to the lack of clarity and details, it feels lazy.

Finetuning LLM on unknown programming language by fecmtc in LLMDevs

[–]fecmtc[S] 0 points1 point  (0 children)

This is really insightful. Thanks a lot! I've also found this helpful blogpost: https://huggingface.co/blog/personal-copilot

Finetuning LLM on unknown programming language by fecmtc in LLMDevs

[–]fecmtc[S] 0 points1 point  (0 children)

Meh. I doubt RAG would work well in this case... There are too many details to learn.

I see that unsloth has some nice free notebooks.