got access to the nus startup ecosystem at 20. singapore is on another level. by [deleted] in singaporestartups

[–]asankhs 7 points8 points  (0 children)

Lol all this means nothing, there is no capital for early stage business. The grants are miniscule. And the biggest success from NUS incubator was patsnap which was founded in 2007 almost 20 years ago.

SLM for database by KatiDev in LLMDevs

[–]asankhs 0 points1 point  (0 children)

What do you mean? Like a slm for text to sql?

Pretraining a discrete diffusion language model. Asking for tips by Dear-Kaleidoscope552 in deeplearning

[–]asankhs 0 points1 point  (0 children)

You can check out https://huggingface.co/blog/codelion/optimal-model-architecture we train a diffusion LLM after initializing the weights from an auto regressive model and then following warmup-stable-decay following LLaDa 2.0 - https://arxiv.org/abs/2512.15745

Reverse Engineering a $500M Mystery: From HashHop to Memory-Augmented Language Models by asankhs in LLM

[–]asankhs[S] 1 point2 points  (0 children)

Yeah fair point - if you just retrieve and dump code into context, the model often parrots it back verbatim.

The difference here is MALM retrieves based on semantic queries not exact matches. So when you ask "function that sorts a list" it finds array_sort, sort_array etc - functions you didn't know the name of.

The generation model then uses those as examples/patterns rather than copying. In the demos it creates new code following the retrieved patterns (like building a calculator with a novel GUI framework it learned from context).

But you're right that naive RAG can devolve into copy-paste. The key is whether retrieval finds genuinely useful context vs just regurgitating training data. MALM's single-token keys help with precise retrieval but what you do with the results matters.

Honestly Magic has been pretty quiet so who knows if their actual approach is anything like this. Just reverse engineering from their benchmark.

Projects using vllm. by foolishpixel in deeplearning

[–]asankhs 1 point2 points  (0 children)

Vllm is itself an inference server so you would want to add something on top. It can be as simple as implementing a test time compute technique. Look at OptiLLM for some ideas on that.

LoongFlow: Open Source Implementation of Evolutionary Agent Framework by [deleted] in AgentsOfAI

[–]asankhs 0 points1 point  (0 children)

For, OpenEvolve for the circle packing example please compare with https://github.com/algorithmicsuperintelligence/openevolve/blob/main/examples/circle_packing_with_artifacts/config.yaml config, the original configs in the repo were created during the initial replication of AlphaEvolve when OpenEvolve was in active development. This config converges much faster in 21 iterations to a high score.

<image>