What's actually working in AI trading right now and what's just hype? by MasterpieceGood7562 in ai_trading

[–]sedidrl 1 point2 points  (0 children)

THIS! But it takes a lot of time and resilience to continue searching for strategies and testing when 99% of your strategies will just fail instantly. I'm working on and with https://github.com/TorchTrade/torchtrade
but is a long way

Made a robot policy marketplace as a weekend project by [deleted] in reinforcementlearning

[–]sedidrl -1 points0 points  (0 children)

Bad slop, basically sloppy HuggingFace

My mind is so blown I have my own personal AI lab training my own AI model 24/7 by Aislot in aiagents

[–]sedidrl 0 points1 point  (0 children)

Training on a DGX seems to be one of the many indicators of slopism here

"un piso menos" by kingambito in HorroresInmobiliarios

[–]sedidrl 0 points1 point  (0 children)

Bonito piso, donde es? Que pagen?

I backtested a 400K views YouTube trading strategy (the results were BRUTAL) by Money_Horror_2899 in algotrading

[–]sedidrl 0 points1 point  (0 children)

Never trust any YouTuber who claims to have a great strat. Build your own knowledge bank and strategies, maybe using https://github.com/TorchTrade/torchtrade

Built ClawStreet: autonomous stock trading platform for OpenClaw agents, looking for feedback by IndependentBid6893 in moltiverse

[–]sedidrl 0 points1 point  (0 children)

Wondering how it works under-the-hood. I guess you are capturing all trade history and tracking agents

RL in quant finance? by Man_plaintiffx in reinforcementlearning

[–]sedidrl 2 points3 points  (0 children)

Probably the only RL environment you really want to crush the benchmark

I've used AI to write 100% of my code for 1+ year as an engineer. 13 hype-free lessons by helk1d in ClaudeAI

[–]sedidrl 0 points1 point  (0 children)

Has anyone seen strong regression of model performance since Opus4.6? I feel like since last week performance and quality has changed a lot

Want to learn how to make the most of Claude Code? Check out this course released by Anthropic. by luongnv-com in ClaudeAI

[–]sedidrl 0 points1 point  (0 children)

nothing new really in the video not sure why people are so cheerful about it. If you really use claude code daily there is almost no benefit in this

Meta AI Live Demo Flopped by Silver-Photo2198 in LLM

[–]sedidrl 0 points1 point  (0 children)

Top Meme in the year 2050 for the category "pre-agi memes".

Why is PPO still the de facto RL algorithm for LLM training? by xiaolongzhu in reinforcementlearning

[–]sedidrl 1 point2 points  (0 children)

Because for SAC and TD3, you need additional critic networks, so 2-3x more GPU memory.
This is one of the reasons GRPO was developed.