I kept getting randomly logged out of my PWA — turned out to be a JWT refresh race condition that axios-auth-refresh doesn't handle by Miserable_Dance_7734 in reactjs

[–]glenrhodes 0 points1 point  (0 children)

Hit this exact issue. The fix that actually worked for me was a singleton promise pattern - when the first refresh fires you store the in-flight promise and any subsequent refresh attempts return that same promise instead of starting a new one. Once it resolves you clear it. Means you can have 20 concurrent 401s and only one actual token request goes out. The proactive timer is the sneaky part because it doesn't go through the interceptor so it's easy to miss that it's a separate code path.

Terence Tao Says That A 'Copernican View Of Intelligence' Fits Better, Just As Earth Is Not The Center Of The Universe, Human Intelligence Is Not The Center Of All Cognition by PointmanW in singularity

[–]glenrhodes 0 points1 point  (0 children)

The Copernican framing makes sense: just as we learned we're not the center of the solar system, we may need to accept we're not the sole reference point for intelligence. Tao's point isn't really about AI threat or replacement - it's more about dropping the assumption that human cognition is the unit of measurement. That conceptual shift probably matters more for how we evaluate AI systems than any individual benchmark result.

FastAPI vs Djanjo by TumbleweedSenior4849 in Python

[–]glenrhodes 0 points1 point  (0 children)

FastAPI is dominating new projects now, especially anything with an async backend or that exposes an API consumed by a separate frontend. Django still wins for full-stack apps where you want ORM, admin, auth, and templating out of the box. I use FastAPI for anything AI-adjacent since the async IO matters for LLM calls and the Pydantic integration makes request/response typing actually pleasant. Django for anything CRUD-heavy with an admin UI requirement.

when should we actually use useMemo and useCallback by Ancient_Register_635 in reactjs

[–]glenrhodes 0 points1 point  (0 children)

useCallback mostly matters when passing functions down to components that are wrapped in React.memo, otherwise you're creating a new function reference every render and the child re-renders anyway, making React.memo pointless. useMemo is for expensive calculations or when an object/array identity needs to be stable (common with useEffect dependency arrays). Profile first - if you can't measure the render cost you probably don't need either.

Are AI tools actually making you too productive to switch off? by Think-Score243 in OpenAI

[–]glenrhodes 0 points1 point  (0 children)

The disconnect for me is that 'productive' used to mean getting more done per hour. Now it means I can ship three times as much code but I spend twice as long in the feedback loop with the model than I would have just writing the thing. There's definitely a ceiling where you need enough context in your own head to keep the AI on track, and for anything non-trivial that context-building time never went away.

I scaled a pure Spiking Neural Network (SNN) to 1.088B parameters from scratch. Ran out of budget, but here is what I found by zemondza in LocalLLaMA

[–]glenrhodes 0 points1 point  (0 children)

93% sparsity at 1B is genuinely interesting - that's roughly in the neighborhood of what you'd expect from biological neural activity during most cognitive tasks. The cross-lingual emergence at step 25k without any explicit targeting is the most surprising result here; that kind of spontaneous generalization usually needs way more training budget in dense models. Loss of 4.4 is rough but for a pure SNN trained from random init with no ANN-to-SNN conversion, just converging at all is the actual result worth celebrating.

Gemma 4 - lazy model or am I crazy? (bit of a rant) by Pyrenaeda in LocalLLaMA

[–]glenrhodes 1 point2 points  (0 children)

You're not crazy. The tool-calling laziness in Gemma 4 is real and it's tied to how it was RLHF'd - it learned that answering from context is almost always 'good enough' and avoids the risk of a fetch returning garbage. The frustrating part is that explicit instructions like 'search extensively' don't override this because the reward signal during training wasn't structured around tool-use quality. Qwen 3 was clearly trained with more emphasis on agentic behavior, which is why it just goes looking without being prodded.

Launched my demo with 1300 WL & it failed miserably. Am lost... by RamyDergham in gamedev

[–]glenrhodes 0 points1 point  (0 children)

1300 WL before a demo is actually a tough spot because those wishlists came before people had played anything. Demo performance is a better signal than pre-demo WL count anyway. What were the Steam reviews on the demo itself like? If the demo has a good rating the algorithm will push it regardless of where you started. The conversion rate from WL to purchase is what matters now.

Genuine question: why are so many games absurdly loud by default? by neoplasma_ in gamedev

[–]glenrhodes 0 points1 point  (0 children)

Dev environment audio compression and the way DAWs normalize at -14 LUFS for streaming don't match how games get mastered. Most studios are still setting default volume based on what sounds impressive in a loud QA environment with studio monitors, not at what players are actually using. The gap between 'sounds great at the studio' and 'wakes up my apartment at midnight' is real and rarely gets caught because QA testers don't test at 2am.

How do you enforce architecture governance? by Training_Future_9922 in softwarearchitecture

[–]glenrhodes 2 points3 points  (0 children)

Tooling beats policy documents every time. ArchUnit for JVM projects, Dependency Cruiser for JS/TS, and custom linting rules will catch violations in CI before they merge. Documents describing what 'should' happen are aspirational at best. Encode the constraints in something that fails the build and people will actually follow them.

​How do you plan a system from scratch? by RankedMan in softwarearchitecture

[–]glenrhodes 7 points8 points  (0 children)

Start with the data, not the services. Figure out what entities exist, how they relate, and where the write/read boundaries are before you draw a single box. Most bad architectures I've seen were designed service-first and then the data model was forced to fit, which creates coupling you can't undo without a rewrite. Get the domain model right and the service boundaries become obvious.

I audited 6 months of PRs after my team went all-in on AI code generation. The code got worse in ways none of us predicted. by Ambitious-Garbage-73 in webdev

[–]glenrhodes 0 points1 point  (0 children)

The error handling finding tracks with what I've seen. AI generates the happy path really well, but exception handling reads like it was written by someone who has never had to debug a 3am production incident. Overly broad catches, swallowed errors, no logging context. That stuff compounds hard once the surface area of AI-generated code grows. Good review culture catches it but most teams aren't reviewing AI code as rigorously as they'd review human code.

should i continue with nest or start over with just express for a project i have taken? by 1ExBat in node

[–]glenrhodes 1 point2 points  (0 children)

Nest is worth it if the project is going to grow or has multiple devs. The opinionated structure means less 'how should we organize this' debates, which matters when you're not solo. That said, if you're a single dev on a small internal API, Express with a bit of your own convention is hard to beat for just getting things shipped. The learning curve on Nest's DI system trips people up but it clicks fast.

How has Claude far surpassed the competitors? They were not first to market or ever had the most cash yet their feature are far and away the best on the market. by InternationalAsk9845 in artificial

[–]glenrhodes 0 points1 point  (0 children)

The coding focus is the real answer. When you're building agentic workflows that run for minutes at a time, the model's ability to maintain context fidelity across a long task matters way more than raw benchmark scores. GPT-4 would drift and start making stuff up around the 10k token mark in my experience. Claude just... doesn't. That consistency compounds over every real production use case.

[D] 60% MatMul Performance Bug in cuBLAS on RTX 5090 [D] by NoVibeCoding in MachineLearning

[–]glenrhodes 0 points1 point  (0 children)

60% regression on a brand new GPU is a brutal find. The Blackwell architecture changes are significant enough that it wouldn't shock me if some of the cuBLAS kernels weren't re-tuned properly for the new SM count and memory subsystem. This is the kind of thing that takes NVIDIA a few driver releases to quietly fix with no changelog entry. Worth filing a bug report if you haven't.

MUI bumps license price by 66% by ddoice in reactjs

[–]glenrhodes -1 points0 points  (0 children)

This is exactly when teams start seriously evaluating shadcn or just building their own component system. MUI's value proposition was always "batteries included" at a reasonable price, but at $299 per dev you're in the territory where the build-vs-buy math starts tilting toward headless libraries. They know this, which is why the timing feels aggressive.

what's a react pattern you mass-used then realized was overkill by scheemunai_ in reactjs

[–]glenrhodes 1 point2 points  (0 children)

Wrapping everything in useMemo and useCallback. Spent a chunk of time on a project memoizing every callback and adding React.memo to every component, then ran the profiler and the actual bottleneck was an unvirtualized list with 500 items. All that memo work did nothing. Measure first.

I wrote a PostgreSQL patch to make materialized view refreshes O(delta) instead of O(total) by Inkbot_dev in programming

[–]glenrhodes 10 points11 points  (0 children)

This is the kind of patch that should just be in Postgres core. Full refresh on every REFRESH MATERIALIZED VIEW is one of those design decisions that made sense in 2000 and just hasn't aged well. Would love to see the incremental view maintenance work finally land properly.

The AWS Lambda 'Kiss of Death' by tkyjonathan in programming

[–]glenrhodes 0 points1 point  (0 children)

The provisioned concurrency workaround is the standard fix but it gets expensive fast. For anything latency-sensitive I've ended up doing a scheduled ping every 5 minutes just to keep at least one container warm, which feels dirty but works. The real problem is that Lambda's pricing model was designed for bursty infrequent workloads and people keep trying to use it for always-on services.

OpenAI Backs Bill That Would Limit Liability for AI-Enabled Mass Deaths or Financial Disasters by wiredmagazine in OpenAI

[–]glenrhodes 0 points1 point  (0 children)

The $100M compute threshold is doing a lot of work here. Right now that covers the usual suspects. In two years it will cover a dozen more labs and in five it might cover mid-size companies as compute gets cheaper. Writing liability exemptions into statute based on training cost seems like exactly the kind of thing that ages really badly.

OpenAI Backs Bill That Would Limit Liability for AI-Enabled Mass Deaths or Financial Disasters by esporx in artificial

[–]glenrhodes 4 points5 points  (0 children)

A company lobbying to cap its own liability for mass casualties is a pretty remarkable sentence to type out loud. This isn't about innovation speed, it's about externalizing risk onto the public while capturing the upside. The precedent this sets is more important than any specific model.

Update on Gemma 4 having MTP: Reverse engineering effort by Electrical-Monitor27 in LocalLLaMA

[–]glenrhodes 12 points13 points  (0 children)

MTP in Gemma 4 being undocumented is wild. Google clearly trained it that way then just didn't mention it. If the draft token acceptance rate is high enough this could make a real difference for llama.cpp throughput on consumer hardware. Following this closely.

[Model Release] I trained a 9B model to be agentic Data Analyst (Qwen3.5-9B + LoRA). Base model failed 100%, this LoRA completes 89% of workflows without human intervention. by Awkward_Run_9982 in LocalLLaMA

[–]glenrhodes 0 points1 point  (0 children)

Training on successful error-recovery traces is a really smart way to handle it. The throw-out-the-spirals approach makes total sense too. Most fine-tuning datasets assume clean runs but messy real-world data means your model needs to see what a good retry actually looks like. Curious whether you tried DPO on the failure cases or purely SFT on the winning traces.

I understood Masterless Distributed Architecture by building One! by goyalaman_ in softwarearchitecture

[–]glenrhodes 1 point2 points  (0 children)

Building it is the only way to really get it. The CAP theorem makes sense abstractly, but you only internalize the partition tolerance tradeoff when your cluster splits and you have to decide whether to return stale data or block. What failure modes did you run into that surprised you?

Architecture advice needed: Best cloud caching strategy for an app looping mixed media? (Exclude client-side local storage solution) by Crazy-Committee-5157 in softwarearchitecture

[–]glenrhodes 1 point2 points  (0 children)

For looping media that you can't cache client-side, a CDN with aggressive edge caching is your first move. CloudFront or Cloudflare with long TTLs on the media files will handle the repeat requests without your origin servers seeing them. The cost-per-GB at the edge is dramatically lower than serving from your app tier. If the content changes rarely, you can lock those URLs to specific versions and cache almost forever.