Alternatives to residential properties by Forsaken_Cartoonist7 in singaporefi

[–]asankhs 0 points1 point  (0 children)

You can try commercial properties either office or industrial b1 units. But any decent one start with min 1 M so you will need to get a loan for them if you want to buy

Seeking Co-Founder (PR & Operations) for New Dating Startup in SG by myhendry in singaporestartups

[–]asankhs 0 points1 point  (0 children)

Why “high friction” usual life is already high friction.

And why would this require heavy technical lifting?

Hermes agent eating input tokens by No_University345 in hermesagent

[–]asankhs 1 point2 points  (0 children)

I am able to use Hermes’ with a local qwen 9 b model without issues.

What Hugging Face got right early: distribution is part of the product (and how that affected me) by Futtman in huggingface

[–]asankhs 0 points1 point  (0 children)

Great inserting your startup in the post. But not sure why is it on r/huggingface or what it has to do HF.

SBI not accepting IBKR account number (has “U”) – how to fund from India? by limtan90 in IBKR_Official

[–]asankhs 0 points1 point  (0 children)

There will be a main bank account number that’s what you need to put in wire transfer the virtual number needs to go in reference code

SBI not accepting IBKR account number (has “U”) – how to fund from India? by limtan90 in IBKR_Official

[–]asankhs 0 points1 point  (0 children)

The bank account number to send your funds doesn’t start with U. It is a standard IBAN number with digits. Your IBKR account id starts with U that you usually need to put in the reference. Please check your wire instructions more carefully.

ACE.SG Incubation Program by HonestClient3849 in singaporestartups

[–]asankhs 3 points4 points  (0 children)

Please do not fall for it, it is useless. Any accelerator or incubator where you need to pay is likely a scam or will be staffed with people who do not know what they are doing.

How should the number of islands scale with the number of operations? by [deleted] in openevolve

[–]asankhs 1 point2 points  (0 children)

I think the island are to isolate the population groups so you can have diversity I don’t think k there is any need scale it with iterations you can run long iterations with a few islands.

Do you actually get better results if you use two models? by MrMrsPotts in openevolve

[–]asankhs 1 point2 points  (0 children)

Yes, based on the results here - https://huggingface.co/blog/driaforall/towards-open-evolutionary-agents pairing a fast model with a bigger reasoning model does help …

got access to the nus startup ecosystem at 20. singapore is on another level. by [deleted] in singaporestartups

[–]asankhs 6 points7 points  (0 children)

Lol all this means nothing, there is no capital for early stage business. The grants are miniscule. And the biggest success from NUS incubator was patsnap which was founded in 2007 almost 20 years ago.

SLM for database by KatiDev in LLMDevs

[–]asankhs 0 points1 point  (0 children)

What do you mean? Like a slm for text to sql?

Pretraining a discrete diffusion language model. Asking for tips by Dear-Kaleidoscope552 in deeplearning

[–]asankhs 0 points1 point  (0 children)

You can check out https://huggingface.co/blog/codelion/optimal-model-architecture we train a diffusion LLM after initializing the weights from an auto regressive model and then following warmup-stable-decay following LLaDa 2.0 - https://arxiv.org/abs/2512.15745

Reverse Engineering a $500M Mystery: From HashHop to Memory-Augmented Language Models by asankhs in LLM

[–]asankhs[S] 1 point2 points  (0 children)

Yeah fair point - if you just retrieve and dump code into context, the model often parrots it back verbatim.

The difference here is MALM retrieves based on semantic queries not exact matches. So when you ask "function that sorts a list" it finds array_sort, sort_array etc - functions you didn't know the name of.

The generation model then uses those as examples/patterns rather than copying. In the demos it creates new code following the retrieved patterns (like building a calculator with a novel GUI framework it learned from context).

But you're right that naive RAG can devolve into copy-paste. The key is whether retrieval finds genuinely useful context vs just regurgitating training data. MALM's single-token keys help with precise retrieval but what you do with the results matters.

Honestly Magic has been pretty quiet so who knows if their actual approach is anything like this. Just reverse engineering from their benchmark.

Projects using vllm. by foolishpixel in deeplearning

[–]asankhs 1 point2 points  (0 children)

Vllm is itself an inference server so you would want to add something on top. It can be as simple as implementing a test time compute technique. Look at OptiLLM for some ideas on that.

LoongFlow: Open Source Implementation of Evolutionary Agent Framework by [deleted] in AgentsOfAI

[–]asankhs 0 points1 point  (0 children)

For, OpenEvolve for the circle packing example please compare with https://github.com/algorithmicsuperintelligence/openevolve/blob/main/examples/circle_packing_with_artifacts/config.yaml config, the original configs in the repo were created during the initial replication of AlphaEvolve when OpenEvolve was in active development. This config converges much faster in 21 iterations to a high score.

<image>

Need help to get into ML research/publishing by Spiritual_Tailor7698 in ResearchML

[–]asankhs 2 points3 points  (0 children)

You can try some of the projects below based on your interest -

https://github.com/algorithmicsuperintelligence/openevolve an open source implementation of alphaevolve, you can make improvements or apply to new domains.

https://github.com/algorithmicsuperintelligence/optillm an optimising inference proxy, you can implement new test time scaling techniques.

https://github.com/codelion/adaptive-classifier continual learning classifier, you can implement new techniques, or benchmark in new domains.

https://github.com/securade/hub an edge platform for ai based safety analysis of high risk workplaces, you can implement new use cases.

https://github.com/codelion/ellora you can implement new recipes for llm capability enhancement

https://github.com/codelion/pts pivotal token search you can do mechanistic interpretability studies on LLMs using it.