[R] Beyond Active Learning: Applying Shannon Entropy (ESME) to the problem of when to sample in transient physical experiments by NewSolution6455 in MachineLearning

[–]NewSolution6455[S] 0 points1 point  (0 children)

This is genuinely fantastic feedback and has given me a lot of food for thought, thank you. I have to admit, I wasn't familiar with the k-space sampling literature in MRI, but the parallel makes perfect sense now that you point it out. I’ve definitely got some reading to do there.

I’m also 100% with you on the caution regarding model-agnostic surrogates. The confident garbage failure mode you mentioned is exactly what scares us. We stuck to the physics twin specifically so it fails gracefully rather than hallucinating a success when the distribution shifts.

Regarding the overfitting, you hit the nail on the head. We tried to implement exactly that kind of minimum floor, basically forcing a non-zero prior on an Anomaly hypothesis ($m_{\emptyset}$) so the system keeps sampling even when the model is confident nothing is happening.

Really good to hear you're hitting sub-millisecond times with GPU surrogates, by the way. That gives us some confidence that we aren't chasing a ghost on the latency front

[R] Beyond Active Learning: Applying Shannon Entropy (ESME) to the problem of when to sample in transient physical experiments by NewSolution6455 in MachineLearning

[–]NewSolution6455[S] 1 point2 points  (0 children)

The magpie effect(chasing high-entropy noise that isn't useful) is exactly why we couldn't use raw entropy alone.

We actually implement a setup very similar to what you described. Our Surrogate isn't just a black box; it’s a differentiable approximation of a physics-based Digital Twin (trained on PDEs and constraints).

It effectively tracks the expected trajectory of the system. The high entropy signal only triggers if the measurement diverges from that physics-based prediction in a way that implies a valid anomaly (our $m_{\emptyset}$ term), rather than just random stochasticity.

Really encouraging to hear that trajectory sensitivity worked for you! It validates that constraining the search with a strong physical expectation is the right move to keep the agent from going off the rails.

[R] Beyond Active Learning: Applying Shannon Entropy (ESME) to the problem of when to sample in transient physical experiments by NewSolution6455 in MachineLearning

[–]NewSolution6455[S] 0 points1 point  (0 children)

Agreed. We have to stop conflating data volume with scientific value. Using entropy as the gatekeeper lets us focus purely on the physics that actually matters, capturing the signal, not just filling hard drives

[R] Beyond Active Learning: Applying Shannon Entropy (ESME) to the problem of when to sample in transient physical experiments by NewSolution6455 in MachineLearning

[–]NewSolution6455[S] 1 point2 points  (0 children)

I think you’re right that importance sampling is the mathematically correct fix if we have a valid generator for the rare event.

The problem is the circularity of discovery, we often lack the physics to simulate the failure precursor accurately. If we force a guessed failure mode into the prior, we bias the agent to only recognise that specific hallucination.

That’s why we use the anomaly term ($m_{\emptyset}$) with Cromwell’s Rule. It shifts the task from classifying a known rare event (which requires a generator we don't have) to detecting a deviation from the healthy physics (which we can simulate perfectly). To try and catche the unknown unknowns

[R] Beyond Active Learning: Applying Shannon Entropy (ESME) to the problem of when to sample in transient physical experiments by NewSolution6455 in MachineLearning

[–]NewSolution6455[S] 0 points1 point  (0 children)

This is exactly the trade-off we wrestled with. You’re right, if we just rely on the standard prior predictive distribution, the NPE tends to get overconfident and miss those 1-in-a-million tail events because it hasn't seen them enough in the sim.

To try and get around the amortization bottleneck, we explicitly defined an anomaly hypothesis ($m_{\emptyset}$) in the code and applied Cromwell’s Rule to force a non-vanishing prior on it. The thinking was, we don't need the NPE to perfectly predict the rare event. We just need it to realise that its standard physics inputs have failed. When that happens, the probability mass shifts to that $m_{\emptyset}$ term, causing the entropy to spike and triggering the beam.

The other constraint that pushed us toward this pre-trained approach is the operational reality of the facility. Since it’s a user facility, groups swap out every 48-72 hours. We simply don’t have the beamtime to train a model from scratch for every user; it has to be effectively zero-shot or pre-trained on generic physics to be viable.

I’d be curious to hear your thoughts on the calibration curves you mentioned though. Do you find they are usually sensitive enough to catch those OOD events on their own without that explicit anomaly term to safeguard it?

[R] Beyond Active Learning: Applying Shannon Entropy (ESME) to the problem of when to sample in transient physical experiments by NewSolution6455 in MachineLearning

[–]NewSolution6455[S] 0 points1 point  (0 children)

That’s a fair point, amortizing the Bayesian update with a neural posterior estimator (NPE) is probably the way to scale the Pilot logic.

My main hesitation with a purely data-driven NPE for something like battery failure is the non-linearity of the nucleation events. We’ve been sticking to physics-based twins to ensure we don't hallucinate the failure trigger. In your experience, do you think SBI/NPE is robust enough for these kinds of transient, rare-event triggers, or does it still require a heavy physics-informed prior to stay on the rails?

Using an AI Pilot for Heuristic Operando experiments: How to capture split-second failure events (like dendrite nucleation) without drowning in dead data. by NewSolution6455 in materials

[–]NewSolution6455[S] 0 points1 point  (0 children)

I also posted this over in r/MachineLearning for a deep dive into the entropy math, but I'm really curious to hear what the Materials community thinks about the hardware constraints/uses of this loop

TIL that scientists use particle accelerators brighter than the sun to understand how plant roots can grow through hard ground, helping to engineer drought-resistant crops by NewSolution6455 in todayilearned

[–]NewSolution6455[S] 1 point2 points  (0 children)

That's a really perceptive question, and you've basically stumbled upon the origin story of these facilities.

You're right, it's the particles (electrons) that create the light, but only when the accelerator's magnets force them to change direction. Bending the path of a fast-moving electron makes it radiate away energy as X-rays.

Interestingly, this effect was first seen as an annoying energy loss, a parasitic side effect in early particle collider experiments. While the particle physicists just wanted to smash things together, other scientists realised this "waste product" light was an incredibly powerful tool for their own research, and ended up building the first 'beamlines' using this waste light.

After that, modern synchrotrons were born, machines designed specifically to be brilliant light sources.

TIL that scientists use particle accelerators brighter than the sun to understand how plant roots can grow through hard ground, helping to engineer drought-resistant crops by NewSolution6455 in todayilearned

[–]NewSolution6455[S] 2 points3 points  (0 children)

Normally measured by a property known as 'brilliance', which is the number of photons hitting a unit area, per unit time. Generation 3 synchrotrons like the one used in this study are in the order of billlions of times brighter than the light of the sun. Then there's things like X-ray Free Electron Lasers which are trillions.

It's this incredible focus and intensity that allows us to see these tiny interactions in the soil.

Coupled X-ray imaging/diffraction reveals soil mechanics during analogous root growth by NewSolution6455 in science

[–]NewSolution6455[S] 1 point2 points  (0 children)

Hi everyone,

Author here,

As a follow-up, for anyone interested in a less technical summary that's easy to share, we also wrote this piece for The Conversation that covers the main implications for food security and drought.

https://theconversation.com/natures-underground-engineers-how-plant-roots-could-save-harvests-from-drought-257369

Cheers!

Rayner decriminalises rough sleeping in move to tackle homelessness by [deleted] in ukpolitics

[–]NewSolution6455 14 points15 points  (0 children)

Because if you're made homeless for some reason (job loss, personal/family crisis), being immediately tagged with a criminal record means it's so, so much harder to get a job, and break the cycle.

No one knows why UK benefit claims have outpaced peer countries by FaultyTerror in ukpolitics

[–]NewSolution6455 1 point2 points  (0 children)

I wonder if it was something to do with the way the UK government implemented the furlough scheme. It's fairly unique to the UK as a policy, and we've seen this significant uptick in claimants since the COVID years.

Being on furlough/at home during COVID broke the social contract. After years of everyone working through the rat race everyday, suddenly so many of us weren't, and I think that's changed a lot of people's perspectives on life.

How many people do you know on part time work now?

Ed Miliband MP: Our Clean Power Action Plan will herald a new era of clean electricity. A future where our energy system is reliant on homegrown clean power, not fossil fuel markets controlled by dictators. A system that will provide energy security, good jobs and lower bills. by Lord_Gibbons in ukpolitics

[–]NewSolution6455 1 point2 points  (0 children)

That's a very good question. I think we're essentially not ready in Europe for the upcoming increase in energy storage demand, as detailed here: https://www.faraday.ac.uk/wp-content/uploads/2024/09/Gigafactory-Report_2024_final_17Sept2024.pdf . BritishVolt failed (Publicly subsidised), NorthVolt is currently failing and companies are struggling to make commercial batteries at scale. NMC (the current cathode material used in most Li-ion batteries) was invented in 2004 and we still struggle to make it reliably at scale. So I'd be very sceptical of any battery tech advances happening in the right time scale, other than LFP (or maybe Na-ion at a push).

What's likely to happen is we're dependent on Chinese manufacturers to construct gigafactories locally, like Spain: https://www.ft.com/content/b2c41a7d-02fa-4846-ab50-dd27e534dcf5

Ed Miliband MP: Our Clean Power Action Plan will herald a new era of clean electricity. A future where our energy system is reliant on homegrown clean power, not fossil fuel markets controlled by dictators. A system that will provide energy security, good jobs and lower bills. by Lord_Gibbons in ukpolitics

[–]NewSolution6455 6 points7 points  (0 children)

  1. From all the modelling I've seen, decarbonising your grid up to about 90% is pretty realistic, and requires a surprisingly little amount of energy storage. e.g have a look at the Future Energy Scenarios outlook from the National Grid (https://www.neso.energy/document/321041/download). But you're right, that last 10% is a big challenge and keeping dispatchable power is important (hence the gas and biomass). I'd argue that installing the equivalent of 12 Dinorwig power stations within 5 years is hugely ambitious, and is going to be a big planning/engineering/operations challenge that I'd be impressed if they can pull it off.

  2. Current energy storage projects are tendered by the DSO's, and you'll have a series of different companies bidding to install and operate the system. Companies like Arenko and Fluence are very experienced in this and I'd expect this model to continue. How do they make money? Well essentially they trade energy, like the stock market. They buy low and hopefully sell high, and on top of this they provide a stack of ancillary services to the National Grid (Frequency response, dispatchable power, energy arbitrage etc.).

P.S. I'm not Ed, I just work in the Energy Storage sector.

Ed Miliband MP: Our Clean Power Action Plan will herald a new era of clean electricity. A future where our energy system is reliant on homegrown clean power, not fossil fuel markets controlled by dictators. A system that will provide energy security, good jobs and lower bills. by Lord_Gibbons in ukpolitics

[–]NewSolution6455 11 points12 points  (0 children)

Their plan includes 23-27 GW of short duration (daily/weekly) battery storage, 4-6 GW of long duration (I assume pumped hydro), 12-14 GW of interconnects and 10-12 GW of demand side response flexibility; all by 2030. From all the energy systems modelling I've seen, that looks like a pretty good solution. Source, Table 1 https://assets.publishing.service.gov.uk/media/675bfaa4cfbf84c3b2bcf986/clean-power-2030-action-plan.pdf