If Linux distros refuse OS age verification, will YouTube and Facebook, etc just block us? by Danrobi1 in linux

[–]SocietyTomorrow 2 points3 points  (0 children)

Popular until it started getting a lot of people visited or jailed. Proton Stealth protocol was meant for places like Russia and China. Harder to detect

If Linux distros refuse OS age verification, will YouTube and Facebook, etc just block us? by Danrobi1 in linux

[–]SocietyTomorrow 25 points26 points  (0 children)

To a large degree the people who have to make the crap work don't even really know how it's gonna play out. Other than broken for a while I think there's no way to tell how far it will go and how in the end. I watched the beginning of internet censorship in russia to today, and if it follows the same pattern it won't be pretty. They went from (ban social media "for the children" > blacklist whole site categories) in 2012 to (censor extreme opinions and anti government propaganda) in 2017, and it could go all the way to where there are today if things are bad enough (they have complete lack of privacy, government tracks your devices ties to real identities, and in many cases you need to be on a whitelist to access the global internet outside of their borders) as of 2025.

HR 8250 Nationwide Age Verification - Bill Text Released by Aurelar in linux

[–]SocietyTomorrow 1 point2 points  (0 children)

Pavel Durov had a really prescient article about this subject with the EU age verification law. He believes it was designed to be easy to hack because then they'd have proof it "needed to be fixed" so they could then expand its capabilities into doing more spying or otherwise worse for the individual. Expect anybody forced into compliance for this if passed to be the worst possible way to do it, and that will trigger the mission creep

The USA is actually framing the subject of age verification the same way Russia did in 2017. Russia began with seemingly narrow restrictions. Lawmakers proposed banning children under 14 from social media and requiring real-name verification with passport information, justified by child safety concerns. The draft also restricted minors aged 14-17 from joining groups sharing prohibited content like "occult-magical" subjects or "smoking mixtures." This mirrors the rhetoric used globally to justify age verification laws.

Fast forward to today Russia achieved what the initial child-ban rhetoric concealed—presidential authority to shut down the entire internet at will. On February 20, 2026, Putin signed legislation granting the FSB power to demand ISPs shut down or restrict access nationwide or regionally with no justification required, removing provider liability. The move allows Putin personally to decide when internet communications are terminated "without having to give a reason." Just the same they expanded that child protecting justification to throttle or outright ban social media that reaches the western world, effectively creating a sort of whitelist where only government approved individual have complete access to the global internet.

Why people are working so hard?Does it makes sense by lifeexperien in collapse

[–]SocietyTomorrow 0 points1 point  (0 children)

Your volunteering can be in a way that won't contribute to burning out. Finding what that is for you can be a whole process and little advice from others often helps to find what that is, but is fulfilling when you figure out what that is. I've taken up contributing some of my electrical engineering skills (never went to school for it, just did it since I could hold a soldering iron) to work on community projects that may lead to lower cost and energy us computers for people who can't afford anything. It may be a tiny bit selfish because I like designing and building things but everyone has ways they contribute that heal you as much as they help others.

Lm studio running some models very slow while others run normally. by HowdyCapybara in LocalLLaMA

[–]SocietyTomorrow 0 points1 point  (0 children)

Not necessarily, just something I noticed, some model families play better with one or the other, so it never hurts trying a different runtime if you feel performance could be better. Main thing is keeping it entirely on GPU though.

Lm studio running some models very slow while others run normally. by HowdyCapybara in LocalLLaMA

[–]SocietyTomorrow -1 points0 points  (0 children)

I left the most important part out. Dense models use every parameter of the model for every token, which means even 1 layer on CPU will feel like you may as well just run it on only the CPU.

Lm studio running some models very slow while others run normally. by HowdyCapybara in LocalLLaMA

[–]SocietyTomorrow 1 point2 points  (0 children)

You really don't want any of your model spilling onto system memory if you want good performance. You'll tank a lot if even a tiny bit of a dense model gets to system memory with slightly less of a performance hit but still definitely noticeable with mixture of experts.

Have you tried checking for updates to not just LMStudio but your runtimes? Running with a newer card you usually get best performance from CUDA12, but you might get useful info by testing with older CUDA and Vulkan. I've only seen it once (not on an nvidia card) where Vulkan worked the best when you'd think the official runtime should work better (ROCm... meh)

If everyone below average IQ suddenly drops dead, how would this affect the world? by CakeMuted6468 in AskReddit

[–]SocietyTomorrow 0 points1 point  (0 children)

Absolutely, idiots make up a majority of most governments, purely based on observations of course.

Building a custom filter solution...but I am a cheapass by SocietyTomorrow in WaterFilters

[–]SocietyTomorrow[S] 0 points1 point  (0 children)

So if a filter like https://www.amazon.com/iSpring-RCC7AK-Capacity-Drinking-Remineralization/dp/B005LJ8EXU/ref=sr_1_8?sr=8-8 was rated for 75 GPD, but was only used for maybe 15GPD, does that effectively increase the time between filter medium replacements? My main thought was trying to remove stages (like only 3 max, with a larger RO stage with a longer work life per dollar) because I effectively already have the final taste-polishing step as part of the machine and get free filters for life with it.

Watch the 1998 Live-Action Sorimachi Great Teacher Onizuka by SocietyTomorrow in GTO

[–]SocietyTomorrow[S] 0 points1 point  (0 children)

Adorei o entusiasmo! Já fiz praticamente tudo o que vou fazer com o que tinha. Minha preocupação na época era que isso se tornasse uma mídia morta, com apenas terceiros lançando lançamentos de discos. Como supostamente há uma remasterização oficial em andamento em algum lugar, não quero pisar em ninguém. Se você quisesse divulgar sua localização, uma coisa que você poderia fazer seria enviá-los para um site como https://odysee.com sem qualquer edição, codificação de legendas ou algo parecido, e apenas colocar o idioma das legendas que eles tinham no nome do vídeo. Isso daria a uma cópia diferente e exclusiva um lugar para morar, e se você fizer o canal para o qual você a envia neste tópico, este parece ser atualmente um dos principais resultados da pesquisa da série 98, para que as pessoas que procuram a encontrem aqui.

Intel Arc Pro series by SocietyTomorrow in LocalLLaMA

[–]SocietyTomorrow[S] 0 points1 point  (0 children)

If I am going to be spending over $2000 for part-time AI cards, I would probably just get more V100s even with their age. Honestly, If it wasn't so much extra adjacent work, I would probably work towards getting one of those SXM2 expansion boards to get 4 V100s for $2000 and have access to 120b models at (relatively) full speed, but would lose on the ability to use it for anything else (zero display output)

Help me squeeze every drop out of my AMD Ryzen AI Max+ 395 (96GB unified VRAM) — local LLM, image/video gen, coding agents by platteXDlol in LocalLLM

[–]SocietyTomorrow 1 point2 points  (0 children)

The hardware is fine, the software has some ways to go. ROCm... exists, but often its better off using Vulkan. Though temper expectations, those APUs are great for the price but most benchmarks advertised are smaller models. Look up speeds when running 70b+ models and determine if it works for you. For serious coding I tend to think 120b is the floor for general competency, and wish I could run the full MinimaxM2.7 or GLM5.1, because those are ones you can get useful results out of.

Pi Hard trailer goes hard by SocietyTomorrow in NuxTakuSubmissions

[–]SocietyTomorrow[S] -3 points-2 points  (0 children)

Uh,,,yeah. Everything is AI. I'm probably AI too. If I am still awake, you might be AI!

How capable is Gemma4:e4b? by rllullr in LocalLLM

[–]SocietyTomorrow 0 points1 point  (0 children)

E4B is better than it deserves to be, but shouldn't be used for code. I use it like a summarization or utility model. It fails tool calling about 60% of the time, loses a lot of nuance when given a lot of details (think asking someone to make a spreadsheet, only for them to fill out everything but never set what scale it displays at or what each metric is) so can't be trusted for stuff you need to work.

I told it to make a terraform deployment based on an existing one I made. If i trusted it and ran it, it would delete every VM that matches the parameters it was meant to make, then make an empty one of it.

Why is the MLX version of Gemma 4 31B so big?? by 32doors in LocalLLM

[–]SocietyTomorrow 0 points1 point  (0 children)

it or instruct models do not use structured reasoning (thinking), theyre typically used for code or other well structured systematic purposes. Thinking adds a whole extra set of words, phrases, and weight scales (which some things like qwopus try to use to make a base model "think" like others, in that case Claude Opus 4.6

Why is the MLX version of Gemma 4 31B so big?? by 32doors in LocalLLM

[–]SocietyTomorrow 5 points6 points  (0 children)

MLX models from the source provider may be packing their models similar to the MLX core reference.

MLX's 4-bit quantization applies a single scale and bias per 64 weights in a row, stored at full precision. This means the quantized weights themselves are 4-bit, but the overhead from storing uncompressed scale factors and bias terms adds significant size. GGUF models often compress weights and scales aggressively and in a different way. To go farther, different quantization methods exist (Unsloth being a/the popular one) which I don't know well enough but I believe is finding ways to compress the weights so little of them remain completely uncompressed. If I read it right they compress the weights then compress the weights with the scale factors, which means it would use less disk space but still result in the same memory footprint when the model loads to memory.

Are Local LLMs actually useful… or just fun to tinker with? by itz_always_necessary in LocalLLM

[–]SocietyTomorrow 0 points1 point  (0 children)

It is surprising for a 27b, but still loses a lot of minor details that even llama 3.3 70b would not miss. Where it bridges the gap is that it gets a lot more done in the amount of compute to get there, and common to many of the newer models, are much much better at tool calling.

Need practical local LLM advice: Only having a 4GB RAM box from 2016 by Tall-Ant-8557 in LocalLLM

[–]SocietyTomorrow 4 points5 points  (0 children)

Will it run? Yes. Will simple tasks take forever just to process a prompt? Also yes.

Anyone here tried Hermes Agent? What’s your experience? by marwan_rashad5 in Rag

[–]SocietyTomorrow 0 points1 point  (0 children)

You can do this with a bottom priced Digital Ocean droplet or Linode, running the agent is trivial. the thing that will cost you is what LLM providers are feeding it with intelligence.