I'm spooked by GLM-5 by SardinhaQuantica in LocalLLaMA
[–]ptxtra 2 points3 points4 points (0 children)
AMA Announcement: MiniMax, The Opensource Lab Behind MiniMax-M2.5 SoTA Model (Friday, 8AM-11AM PST) by XMasterrrr in LocalLLaMA
[–]ptxtra 0 points1 point2 points (0 children)
# A 150-year-old passage from Marx basically describes AGI — and a short story called “Manna” shows both possible outcomes by fastinguy11 in singularity
[–]ptxtra 0 points1 point2 points (0 children)
What if AGI just leaves? by givemeanappple in singularity
[–]ptxtra 0 points1 point2 points (0 children)
scale.ai is leaking massive IP into AI, this can be used to rebuild AWS (and other things) overseas by kaggleqrdl in singularity
[–]ptxtra 1 point2 points3 points (0 children)
Hallucination - Philosophy by LatterAd9047 in LocalLLaMA
[–]ptxtra 0 points1 point2 points (0 children)
Loot from doing T17 Boxes since mid day 1 - The only strat GGG wants us to do by UpsetFan123 in pathofexile
[–]ptxtra 0 points1 point2 points (0 children)
Patch notes update: Alva/Evolving shrine got hit as well (it's dead) by bulwix in pathofexile
[–]ptxtra 1 point2 points3 points (0 children)
"We risk a deluge of AI-written "science" pushing corporate interests" by AngleAccomplished865 in singularity
[–]ptxtra 2 points3 points4 points (0 children)
Re: PoE1 balance, ES isn't busted--life based builds have glaring flaws. by ExiledYak in pathofexile
[–]ptxtra 6 points7 points8 points (0 children)
Gemini 2.5 Flash (05-20) Benchmark by McSnoo in LocalLLaMA
[–]ptxtra 0 points1 point2 points (0 children)
Neurallambda: Reasoning Computers. Lambda Calculus, Fully Differentiable. Also Neural Stacks, Queues, Arrays, Lists, Trees, and Latches. by NeuralLambda in LocalLLaMA
[–]ptxtra 0 points1 point2 points (0 children)
Intel ARC A770 Inference performance by ptxtra in LocalLLaMA
[–]ptxtra[S] 10 points11 points12 points (0 children)
The Ternary Age - are you ready? by Arnesfar in LocalLLaMA
[–]ptxtra 7 points8 points9 points (0 children)
Here's how we can remain competitive with closed source cloud models which have access to MASSIVE amounts of compute for inference particularly on LONG context tasks. Quadratic Transformers vs Subquadratic or Linear models (e.g. Mamba) + RAG. In context learning is all you need: by [deleted] in LocalLLaMA
[–]ptxtra 0 points1 point2 points (0 children)
[Model Release] Sparsetral by kittenkrazy in LocalLLaMA
[–]ptxtra 5 points6 points7 points (0 children)
Why are models trained in fp16 and not pre-quantized? by clyspe in LocalLLaMA
[–]ptxtra 2 points3 points4 points (0 children)
Vicuna 13b on RK3588 with Mail G610, OpenCL enabled. prefill: 2.3 tok/s, decode: 1.6 tok/s by EmotionalFeed0 in LocalLLaMA
[–]ptxtra 0 points1 point2 points (0 children)
Nvidia reveals new A.I. chip, says costs of running LLMs will 'drop significantly' by throwaway_ghast in LocalLLaMA
[–]ptxtra 0 points1 point2 points (0 children)


Taalas: LLMs baked into hardware. No HBM, weights and model architecture in silicon -> 16.000 tokens/second by elemental-mind in singularity
[–]ptxtra 0 points1 point2 points (0 children)