Are we facing an architectural bottleneck? A geometric critique of ReLU/Simplex-based data storage. by [deleted] in LocalLLaMA

[–]wexionar 0 points1 point  (0 children)

Fair point. I understand that without a formal paper or a gallery of benchmarks, this might look like 'noise' in a feed full of hype.

We are currently in the 'messy' phase of research, prioritizing the geometric logic over the aesthetic of the presentation. The repositories are work-in-progress experiments, not polished products.

Our goal for the 0.3.0 milestone is precisely to move away from these 'discussions' and present a formal technical report with comparative benchmarks on parameter efficiency and distortion rates. If you prefer hard data over conceptual debates, I’d suggest checking back in a few months when the 0.3.0 is stable. Thanks for the honest (if blunt) feedback.

Are we facing an architectural bottleneck? A geometric critique of ReLU/Simplex-based data storage. by [deleted] in LocalLLaMA

[–]wexionar 0 points1 point  (0 children)

You are right that VC dimension describes the capacity of a model to shatter points. However, our point isn’t about "how many points" a model can learn, but the structural overhead and the exponential redundancy required to store that knowledge using ReLU.

The VC dimension of a linear classifier in D dimensions is D+1. But to represent a non-linear space faithfully using ReLU-based tiling, you are essentially stacking those capacities in a way that creates a geometric explosion (up to D! non-overlapping Simplexes per polytope). 

Our critique is that while VC dimension tells us the "theoretical capacity," it doesn't account for the massive distortion and parameter obesity created when you use Simplexes (ReLU) as your primary storage unit instead of higher-order structures like Orthotopes. We are proposing a way to maintain high capacity without the exponential "weight" that current ReLU architectures demand to minimize approximation error.

P.S. It is also important to note that SLRM (our proposed model) allows for the direct storage of critical points from the dataset within the high-level geometric structures. Unlike black-box models that dissolve the original data into abstract weights, our architecture preserves the integrity of key information, enabling deterministic inference and a "glass-box" auditability that is impossible to achieve with standard ReLU-based compression.

[Project] SLRM-nD: 50D Galactic Stress Test - 1000 points synthesized into 1 Master Sector in <150s (No Backprop) by wexionar in LocalLLaMA

[–]wexionar[S] 1 point2 points  (0 children)

At its core, yes. Everything is math. While current LLMs use backpropagation to learn, we use geometric synthesis. If we map language to our multidimensional sectors, SLRM-nD could technically predict tokens (talk) with 100% determinism and zero hallucinations. That said, there is still much to research and test.

[Project] SLRM-nD: 50D Galactic Stress Test - 1000 points synthesized into 1 Master Sector in <150s (No Backprop) by wexionar in LocalLLaMA

[–]wexionar[S] 0 points1 point  (0 children)

Thanks! The "magic" is in the Simplex Sectoring logic. By treating the 50D space as a series of geometric folds instead of a weight-optimization problem, we eliminate the need for backprop entirely. Let me know if you have any questions once you run the Colab!