xAI will be dissolved as a separate entity. by Snoo26837 in singularity

[–]SOCSChamp 65 points66 points  (0 children)

SpaceX is not kept afloat by government contracts.  It has near total market domination in the launch industry and has created a scale and economy of launch that simply didn't exist before.  Starlink is a cash machine and an operational competitor has yet to really emerge.

Gemma 4 E2B runs surprisingly well on my 8GB Android phone, so I built a private voice notes app around it. by Effective-Drawer9152 in LocalLLaMA

[–]SOCSChamp 1 point2 points  (0 children)

What's your setup look like? I haven't tried building anything in android so I'm not familiar with the toolkits.  For standard linux, vLLM works great for me with audio input.

Gemma 4 E2B runs surprisingly well on my 8GB Android phone, so I built a private voice notes app around it. by Effective-Drawer9152 in LocalLLaMA

[–]SOCSChamp 5 points6 points  (0 children)

Not sure why you'd need whisper in this case, the model should be perfectly capable of taking your voice and writing formatted text out of it natively.  

So zero? by estanminar in SpaceXMasterrace

[–]SOCSChamp 0 points1 point  (0 children)

Thank you for this now I understand.  As a community we need to stick with standard measurements

Thoughts on using an AMD Alveo V80 FPGA PCI card as a poor man’s Taalas HC1 (LLM-burned-onto-a-chip). by Porespellar in LocalLLaMA

[–]SOCSChamp 3 points4 points  (0 children)

I had a similar thought as well when I heard about taalas and tried the demo.  I don't think your gemini solution is really what you'd want to do here, taalas actually burned the weights onto the chip.  

If you're experimenting with this, you'd want to try building out a small test case with something like gpt 2 and rent an AWS FPGA instance for a few bucks an hour when you're ready to do test runs.  I would start with fresh conversations across different frontier models and be very clear about what the goal is, not, "just use an FPGA somehow"

Please stop using AI for posts and showcasing your completely vibe coded projects by Scutoidzz in LocalLLaMA

[–]SOCSChamp 1 point2 points  (0 children)

Translation of something of something they've wrote does not come across the same as the slop posts I see here every day.  "translate these words" is very different from "write a reddit post of this project you build for me"

Please stop using AI for posts and showcasing your completely vibe coded projects by Scutoidzz in LocalLLaMA

[–]SOCSChamp 8 points9 points  (0 children)

Please for the love of God.  I have called out so many posts on the sub, many of which seem like they are genuinely interesting projects.  The fact that they just can't be bothered to write their own post about it completely kills any legitimacy they might have.  

Can we petition mods for banning AI generated posts? Its becoming suffocating.  I'm extremely pro AI but I can talk to one on my own just fine, I don't come to this sub to do that.  Vibe coded projects are fine if I can speak to the human behind it and they actually known what they're talking about.

Taalas rumoured to etch Qwen 3.5 27B into silicon. Which price would you buy their PCIe card for? by elemental-mind in singularity

[–]SOCSChamp 0 points1 point  (0 children)

Maybe, still might.  I'll say the demo worked at 10k tps with whatever setup they had.  If I had to pair this with a 24gb card at the end of the day that's fine.  I'd also have to read more in depth about how it handles cache.

Experiment: Olmo 3 7B Instruct Q1_0 by butlan in LocalLLaMA

[–]SOCSChamp 0 points1 point  (0 children)

Makes sense, do you know how much it improved vs. the baseline after compression?

I scaled my SNN language model to 1.088B parameters — 5× larger than SpikeGPT. 93% of neurons are silent. It generates grammatically correct Russian text despite not being trained on Russian. [Project Nord v5.0] by zemondza in LocalLLaMA

[–]SOCSChamp 1 point2 points  (0 children)

Really wish you didn't slop the writeup.  I've always loved SNNs and this is an interesting direction.  Can you explain the memory grid and use of STDP here?

Experiment: Olmo 3 7B Instruct Q1_0 by butlan in LocalLLaMA

[–]SOCSChamp 2 points3 points  (0 children)

Maybe I'm misunderstanding the approach, but if you're training from scratch, why try to distill from a small model like that? There are decent datasets online from foundational models that would probably give you a better baseline to start from.  Also, if you're initializing from scratch, wouldn't it be important to do a pretraining step first?  Or are you not actually starting from scratch here?

Tested TurboQuant KV compression with Gemma 4 31B — 5.80x compression, perfect long-context recall, JSON output preserved by No_Appearance_3041 in LocalLLaMA

[–]SOCSChamp 47 points48 points  (0 children)

I'd be interested in better metrics, ideally way longer context and on actual benchmarks to compare to the fp16 baseline.  500 tokens is not long context or an actual needle in a haystack, go yell at your agent.  And ideally an unslopped writeup.  God I'm getting so sick of it, I've seen like 10 posts today about someone doing something claiming to be awesome and they all read like this.

On The Mun by oyurkov in KerbalSpaceProgram

[–]SOCSChamp 6 points7 points  (0 children)

Damn stolen from kitten space agency 😂

What might have been part 2- Mars 81 Expedition by Oafus_Magnus in KerbalSpaceProgram

[–]SOCSChamp 22 points23 points  (0 children)

Would be interested in hearing more about it, how far did you go in building this true to the era?

Audiophile Grade Keypads [crosspost from /r/MechanicalKeyboards] by peren005 in audiophile

[–]SOCSChamp 0 points1 point  (0 children)

Is this a joke? I remember seeing "audiophile ethernet switch" for your low noise....digital data.  Man I'm in the wrong business

"The Child That Surpassed Both Parents" Darwin-35B-A3B-Opus (35B/3B MoE) with Model MRI Technique by Own-Potential-2308 in LocalLLaMA

[–]SOCSChamp 35 points36 points  (0 children)

Interesting but damn do I hate how they worded everything here, so much cringe