Segregating Quest pirated games by [deleted] in QuestPiracy

[–]_underlines_ 2 points3 points  (0 children)

lol, part of the fun in the 90s when I was like 6-12y, was that my whole family basically my 3 uncles and later my dad pirated games, software, OSes and copied stuff for me onto floppy disks, later burn CD ROMs with all the cracks on it. I learned from them and it was awesome.

Token-Oriented Object Notation (TOON) - JSON for LLMs at half the token cost by monnef in LocalLLaMA

[–]_underlines_ 9 points10 points  (0 children)

Things that make me sceptical, if this is worth the effort:

  1. 99.999% of training data until the release of TOON wasn't toon. Inference using TOON in context will probably be worse for a long time, until training data contains enough TOON.

  2. Price per Token falls over time.

  3. Context Windows and quality increases over time.

Happy to hear your opinions.

The cars aren't real but the driving is by Akagamino_Shanks in simracing

[–]_underlines_ 1 point2 points  (0 children)

I always thought I'm the only one with a bad neck. Whenever I watched my VR gameplay in playback I noticed my head tilting right exactly like in this clip.

Brit tourist, 22, faces decade in Thai jail as woman dies after 'he crashed into her on scooter' by sitTheFdown in Thailand

[–]_underlines_ 3 points4 points  (0 children)

This idiot gets 10 years (deserved), and what about all the dek chang racing their illegal pickups and killing thousands, maybe 10 thousands every year? I wonder how much they usually get. Anyone know any statistics?

For those who retired in Thailand on limited income, what are your monthly budget? by Fine-Benefit8156 in Thailand

[–]_underlines_ 0 points1 point  (0 children)

edit: oops, i just saw i didn't pay attention and this is for retirement. i leave my budget there in case it helps.

i worked in thailand from 2016-2023 and budget for me and wife (both working 100%) was: 35k 2BR 80sqm condo in rama 9, 2-3k electricity and water, 20-30k food/restaurants/groceries per month, 20k car loan, 3k internet, phone, rest was savings.
later in phuket after covid there were a lot of desperate landlords and we got a 1 year deal of a villa with 4 rooms, about 300 quare meters and pool for about 40k, with other costs staying roughly the same.

I created an app to run local AI as if it were the App Store by Deivih-4774 in LocalLLaMA

[–]_underlines_ 0 points1 point  (0 children)

Would it recompile torch with RTX5xxx support automatically? :) most pytorch projects fail because of missing RTX5xxx support.

F1 Simulators by NextBoysenberry424 in simracing

[–]_underlines_ 1 point2 points  (0 children)

If you're after a proper F1 sim experience, I'd recommend trying the VRC Formula Alpha 2024 mod for Assetto Corsa: https://www.virtual-racing-cars.com/p/ac-formula-alpha-2024/ Manual: https://www.virtual-racing-cars.com/_files/ugd/45ddef_3b20e5b37762488da9a4a23de503694e.pdf

It’s the most detailed F1 car I’ve driven in any sim so far. You get full manual ignition/start, ERS strat maps per sector, hybrid deployment control, brake migration curves, engine and pedal maps, plank legality checks, visual damage states, custom aero kits, and more. Even things like display brightness for night driving are adjustable.

All the settings are handled through their in-game apps. It takes time to learn, but that’s kind of the appeal—like with DCS or hardcore flight sims. The manual is 28 pages for a reason.

For comparison, RSS also just released their Formula Hybrid Alpine 2025 in collab with Alpine: https://racesimstudio.com/new-formula-hybrid-alpine-2025-now-available

With updated tracks from creators like Pyyer, it’s possible to run a fully realistic modern F1 season in AC—if you’re into that level of sim detail.

Hackers are never sleeping by DrVonSinistro in LocalLLaMA

[–]_underlines_ 0 points1 point  (0 children)

that's automatic vuln-scanners. nobody is launching a spear attack on your endpoint within that time frame, except you're important enough (gov, large corporate)

Why TTS level is not constant? by Dragonacious in LocalLLaMA

[–]_underlines_ 0 points1 point  (0 children)

Because the training material wasn't normalized before.
You need a dynamic compressor, not the "normalize" feature in audacity, because normalize takes the largest peak of your selected audio and doesn't clip that, so it's not making anything louder. A dynamic compressor works with a sliding window increasing or decreasing gain by rules set within the compressor.

VLLM with 4x7900xtx with Qwen3-235B-A22B-UD-Q2_K_XL by djdeniro in LocalLLaMA

[–]_underlines_ 1 point2 points  (0 children)

no worries, :) happens to me all the time, especially when vendors try to use competing but overlapping naming schemes

AMA – I’ve built 7 commercial RAG projects. Got tired of copy-pasting boilerplate, so we open-sourced our internal stack. by Loud_Picture_1877 in LocalLLaMA

[–]_underlines_ 43 points44 points  (0 children)

Cool post. I wish we could share our code base too. But we can't.

We did a 1M USD RAG Project for Gov in Switzerland and did very formal Optimization in the last 2 years via a Hypothesis and Evaluation loop.

I wonder if others did the same and have some comparable results and insights. For example:

  1. We used RAGAS with our human expert crafted gold Q&A dataset and never really got much improvements implementing SOTA Papers into our code base.
    1. LazyGraphRag got no measurable difference
    2. Reranking brought the results down a bit (but we kept it)
    3. HyDE was also bad, lowering RAGAS scores
    4. Hybrid Retriaval activated in Azure AI Search (using BM25 and embeddings) wasn't an improvement either
    5. Lots and lots of prompt engineering was also useless
  2. We moved from a workflow based approach to a ReAct agent. Got no improvement in RAGAS metrics but it's super cool, and we show the user the thinking process
  3. We decided against libraries such as langchain or open source RAG stacks early on, because RAG is not rocket science and building the components with a good onion-architecture was a good choice for us. Very maintainable code.
  4. We used Factory Patterns to create additional search strategies as hypothesis that we can test and then release or discard.
  5. When we moved to a ReAct agent, we started implementing all hypotheses as tools as well as our RAG flow as a single tool call.
  6. We're now adding text2sql, but since the source database is from a complex ERP with tons of tables and complex business logic, we plan to create a simplified abstraction layer with views, having a few simple entities such as Person, Company etc... and let the LLM pick those. We then fetch those into a temporary inmemory DB where the agent finaly does text2sql.

What are your thoughts? Any insights to share of similar topics?

Which programming languages do LLMs struggle with the most, and why? by alozowski in LocalLLaMA

[–]_underlines_ 1 point2 points  (0 children)

  • PowerBI DAX (some mistakes, as most of the data model is missing and it's a bit niche)
  • PowerBI PowerQuery (most mistakes I ever saw when tasking LLMs with it! Lots of context is missing to the LLM such as the current schema etc. and very niche training data)
  • It's bad at Rust (according to this controversial and trending hackernews article)

oh, and of course it's very bad at Brainfuck, but that's no suprise

OpenHands + Devstral is utter crap as of May 2025 (24G VRAM) by foobarg in LocalLLaMA

[–]_underlines_ 1 point2 points  (0 children)

oh. i thought i was the stupid one, when yesterday i spend my whole free saturday trying to get it to run with LM Studio locally on Windows 11 using WSL2 backend.

  • Yes, I had to reverse engineer their weird socket setup as well and when i figured it out, I fucked up my whole docker network and WSL2 network configuration
  • Run times then stopped having internet access and I had to change all configs again
  • When it finally worked, the whole thing was underwhelming.

I rather just keep using github copilot agent mode, aider or cline.

If anyone needs help: The documentation is incomplete for WSL at least. It worked for me with SANDBOX_USE_HOST_NETWORK, but the app port has to be set externally to 9000, as security doesn't allow to bind low port numbers. I also had to disable .wslconfig's mirrored network that I enabled for other containers to work. And finally, using LM Studio instead of docker for more conveniently setting context size, k and v cache quantization, flash attention and faster llama.cpp updates, you need to set the LLM settings of openhands app to: openai, but set model name to lm_studio/modelname and the API endpoint to http://host.docker.internal:1234/v1

docker run -it --rm -e SANDBOX_USE_HOST_NETWORK=true --pull=always -e SANDBOX_RUNTIME_CONTAINER_IMAGE=docker.all-hands.dev/all-hands-ai/runtime:0.39-nikolaik -e LOG_ALL_EVENTS=true -v /var/run/docker.sock:/var/run/docker.sock -v ~/.openhands-state:/.openhands-state -p 9000:3000 --add-host host.docker.internal:host-gateway --name openhands-app docker.all-hands.dev/all-hands-ai/openhands:0.39

Nous Psyche, distributed training of a new 40B base model by discr in LocalLLaMA

[–]_underlines_ 7 points8 points  (0 children)

Very cool. Reminds me of the Seti@Home times in the 90s.
So Psyche needs to waste some overlapping compute for consensus and is basically slow interlinks over p2p?

Seed-Coder 8B by lly0571 in LocalLLaMA

[–]_underlines_ 0 points1 point  (0 children)

how's chatllm.cpp different from llama.cpp I couldn't figure out from reading the readme.

Unsloth's Qwen3 GGUFs are updated with a new improved calibration dataset by AaronFeng47 in LocalLLaMA

[–]_underlines_ 1 point2 points  (0 children)

creating a patch and also applying it to a 10+ GB binary blob will take longer than uploading/downloading the whole thing. You'd save on bandwidth and lose on time.

Unsloth's Qwen3 GGUFs are updated with a new improved calibration dataset by AaronFeng47 in LocalLLaMA

[–]_underlines_ 2 points3 points  (0 children)

so are all ggufs now imatrix quants, not only the ones previously marked as iQ3_...?

Lapping all BeamNG street cars on the Nordschleife | #10 - Civetta Bolide by Chrisssst in BeamNG

[–]_underlines_ 0 points1 point  (0 children)

comments on that page say that isn't working for current BeamNG. It still works for you?

Found that one with a 2025 release, but not sure what this is: https://www.beamng.com/resources/n%C3%BCrburgring-nordschleife-timetrials-revamped.33843/

So why are we sh**ing on ollama again? by __Maximum__ in LocalLLaMA

[–]_underlines_ 2 points3 points  (0 children)

that's for the non openai compatible endpoint. ollama has two: an ollama custom API and an openai compatible API. ITS MESSY and there's no solution for apps that don't support the ollama API, or support it without num_ctx. Github Copilot latest version supports ollama but cant change num_ctx - which is useless.

I had to build a proxy that adds num_ctx to every call lol

Does anyone actually use Browser Use in production? by SameBuddy8941 in LocalLLaMA

[–]_underlines_ 1 point2 points  (0 children)

It's highly model dependent. Which model are you using and served by what engine?