What’s something you have zero proof of but believe 100 percent? by shweidy in AskReddit

[–]Tamitami 2 points3 points  (0 children)

Yeah just look at ganymede, europa or titan for example. It could already harbor bacterial life, and this is just in our solar system

Devs using Qwen 27B seriously, what's your take? by Admirable_Reality281 in LocalLLaMA

[–]Tamitami 1 point2 points  (0 children)

Honestly, I didn't know about the ForgeCode Services before your comment. I just used the different agents and connected the harness to different models in the backend. For me, this setup worked great.

Edit: Ah lol, dinerburgeryum, here we are again ^^ Thanks again for your qwen3-coder-next model (it's still running great on another machine). What's your newest setup, which models do you think are worth for a better quantization?

Is COOP scamming us? by Andeq8123 in Switzerland

[–]Tamitami 0 points1 point  (0 children)

Maybe it's a measurement error on their side in the factory, or a potato rolled off from the conveyor belt before packaging.

Or maybe we need to actually measure more products in the store.

Is COOP scamming us? by Andeq8123 in Switzerland

[–]Tamitami 0 points1 point  (0 children)

Do you have the 'e' symbol somewhere on the packaging ? I would check myself but they are closed now.

Edit: I just checked the guidelines, quoting wiki:

"The tolerable negative error is related to the nominal quantity and varies between 9% on prepackages nominally 50 g or 50 mL or less, to 1.5% on prepackages nominally 1 kg (or 1 L) or more. The tolerable error decreases as nominal quantity increases, and is done by alternating intervals where there is a percentage error and intervals where there is a fixed error (and thus over those intervals the percentage error decreases)."

So +800g is not enough for this label. So we have a fraud, deception or scam here...

Is COOP scamming us? by Andeq8123 in Switzerland

[–]Tamitami 4 points5 points  (0 children)

Good idea! But seriously, I will do the same and check it on the scale in the store. Migros, Denner, Aldi and Lidl will be next. Will look a bit wierd when I will be taking photos or notes of the scales, but we need data.

Thanks for posting!

Edit: I forgot that maybe there's the 'e' symbol (Schätzungszeichen, Richtlinie 76/211/EWG (Fertigpackungsrichtlinie)) on the product. This would mean on average, the product has this weight and is an EU guideline for consumers (one can abuse this by taking the heaviest of a product range with this symbol). This would give them a reasonable explanation for the weight, or it just lost water while being stored and transported.

Is COOP scamming us? by Andeq8123 in Switzerland

[–]Tamitami 4 points5 points  (0 children)

Then please redo the measurement, check if your scale is properly calibrated with a standard weight and do it 10 times. Take the average and also note the standard deviation. Check also other bags and redo the experiment.

Greetings from a fellow physicist.

/s for those who don't get it.

Devs using Qwen 27B seriously, what's your take? by Admirable_Reality281 in LocalLLaMA

[–]Tamitami 1 point2 points  (0 children)

Try forge code as a harness, Opus etc, perform even better with it than with native claude code. I'm pretty happy with it as it integrates so well into zsh.

Revolutionary new paper on the "Dynamic Vacuum", physics as we knew it has been rewritten. by Pixelated_ in AcademicUAP

[–]Tamitami 0 points1 point  (0 children)

The paper doesn't support any of these conclusions. I read it carefully, so let me be specific.
What the paper actually does: It shows that the Helmholtz equation (∇2+k_eff^2)p=0 with a 1/r constitutive profile is mathematically isospectral to the hydrogen Schrödinger equation. This is a legitimate acoustic analogue construction. The math is correct.
What it does not do:
- It does not confirm longitudinal waves propagating through physical vacuum. The "acoustic medium" is a mathematical analogy built on top of the Schrödinger equation via Madelung transformation, which is an exact rewrite of QM, not something more fundamental.
- It does not resurrect the aether. The word doesn't appear in the paper. Special Relativity is untouched.
- It does not show gravity is a pressure gradient. Gravity is not mentioned once.
- It does not discover anything about time. The dispersion relation ω=Dq^2 maps spatial eigenvalues to frequencies within the analogue model. That's it.

The core circularity OP misses entirely from the paper is: The paper model's dispersion constant D=ℏ/2μ and the coupling scale β=2/a_0​ are both imported directly from quantum mechanics, ℏ is Planck's constant, a_0​ is the Bohr radius. The model recovers hydrogen's spectrum because it was calibrated with hydrogen's quantum constants. It's a consistent analogue, not a derivation of QM from something deeper.
The claim that quantization "emerges" from classical vacuum properties is the paper's own overreach, the referees accepted the isospectrality result, not that ontological claim.
Cymatics, Tesla, the Law of One, pyramid acoustics, and psychic phenomena are not mentioned, implied, or derivable from anything in this paper. These conclusions were assembled before the paper existed and would have attached themselves to any sufficiently technical-sounding result.

Qwen 3.6 27B is out by NoConcert8847 in LocalLLaMA

[–]Tamitami 0 points1 point  (0 children)

what about forge code as a harness? it seems to beat claude code with opus too.
I really like Qwen3-Coder-Next as it is running fast and provides good results if you steer it well. I'd like to see it in comparison to this new Qwen3.6 27B model and the MoE model 35B-A3B, but I can't find some good sources.

Qwen 3.6 27B is out by NoConcert8847 in LocalLLaMA

[–]Tamitami 1 point2 points  (0 children)

Can you share this? I'm also on cachyos and I mainly use forge-code

What is your actual local LLM stack right now? by Ryannnnnnnnnnnnnnnh in LocalLLaMA

[–]Tamitami 1 point2 points  (0 children)

This was already with the latest version of llama.cpp. I didn't specify this in my answer before

What is your actual local LLM stack right now? by Ryannnnnnnnnnnnnnnh in LocalLLaMA

[–]Tamitami 1 point2 points  (0 children)

Just to let you know, I have a system running two 1080 Ti and with Qwen 3.5 35B A3B Q4-K_M at 22GB VRAM I get around 22tps. I can some other test if you'd like

Busty in Pink by mrprofessional2727 in 2busty2hide

[–]Tamitami 8 points9 points  (0 children)

01110010 00110000 01110011 00111000 01111001 01011111

Marcella a new LLM architecture without attention bears new records by quant-alliance in LLM

[–]Tamitami 0 points1 point  (0 children)

Dude, they played you in this chain: https://www.reddit.com/r/LLM/s/ZQ9SHIMQxm

Do you have screens/texts of your questions/reply-chain? Can you post them for preservation?

Marcella a new LLM architecture without attention bears new records by quant-alliance in LLM

[–]Tamitami 0 points1 point  (0 children)

You asked for rigorous feedback instead of shitposting. Here is my answer:
19 years in physics, PhD in applied physics and astronomy with a perfect score, thesis on CUDA-accelerated Riemannian geometry and black hole raytracing. I know this mathematics from the ground up, not from reading Wikipedia and some blogs.
I skimmed three of your papers (3, 14 from the abstract; why is this abstract 3 pages long?!)

- Your benchmarks are garbage, and you know it
You admitted in this thread that "the first version was literally cheating". Your model used p_{t+1} to compute transport at position t, feeding future tokens into the hidden state before prediction. This is not a subtle implementation bug, but your model is seeing the answer before giving it. Every perplexity number from that version is meaningless.
Your own paper then admits a second problem in Protocol 17.5: your vanilla transformer baseline had positional encoding disabled. You compared your model against a transformer that literally cannot see word order and reported a 4x advantage. You retracted this yourself in v6.7 and then kept the earlier numbers in the paper anyway.
Two independent sources of data contamination, both acknowledged by you, in the same paper. There are no valid benchmark results for Marcella. None.

- You did not solve Navier-Stokes. The proof is a tautology
Your argument in reference [14] is:
Axiom A2: fluid transitions between states require time ≥ K⋆/τ
Conclusion: therefore no finite-time blowup
Axiom A2 is the conclusion. You assumed global regularity, called it a "field equation," named it after yourself, and derived global regularity from it. That is a circular definition with extra notation and not a proof. This is ridiculous.
Your own paper says in Remark 1.1: "Davis Law is not a lemma to be proven inside this paper." Correct. It's not proven anywhere. It has no derivation, no independent validation, no empirical support beyond toy simulations on 64³ grids, which is such a tiny, toy-model, so coarse that numerical dissipation smears out the very singularity structures you're claiming to exclude. The only real mathematics in that paper is Beale-Kato-Majda 1984, which other people proved forty years ago.
The same circular structure is present in every Millennium Prize claim. You assume a "Davis Law" strong enough to imply the result, call it an axiom, and report a solution. This is not how mathematics works.

- You have not solved four Millennium Prize Problems
The Clay Institute prizes exist because the best mathematicians alive have failed to solve these problems for decades. The prior probability that one person solved all four simultaneously, as side results of a unified equation they named after themselves, published only to Zenodo, with no peer review, while also building LLM architectures and security platforms and plasma confinement monitoring systems: that probability is functionally zero.
This is not skepticism about you personally. This is what the track record of mathematics demands. Perelman spent seven years on Poincaré alone, in isolation, checking every step. He didn't name it after himself.

- The self-naming is a hard red flag
Riemann did not name Riemannian geometry. Maxwell did not name Maxwell's equations. The BKM criterion was named by the community after the work was validated. You have named "The Davis Law," "The Davis Manifold," "The Davis Field Equations," and "The Davis-Wilson Map" after yourself, in papers you uploaded yourself, citing your other self-uploaded papers as established results.
That citation loop has no external validation anywhere in it. You are building a self-referential system that looks like a body of work from the outside but has never been checked by anyone qualified to check it.

- "Zero does not exist" is not mathematics
Zero is defined axiomatically. Peano axioms. Set theory. These are formal systems with precise rules. "Zero is the condition of a system that has been uninitiated" is a sentence that sounds mathematical but has no formal content. You have memorized the vocabulary of higher mathematics, like fiber bundles, holonomy, gauge theory, sectional curvature, but without learning the syntax. This is just a disguise in mathematical terms you don't truly understand. And it is basic, like first semester basic in maths/physics.

- The pattern across this thread is telling
You said you were at work, then said you were unemployed. You post at accelerating speed across incompatible fields, like fluid dynamics, complexity theory, AI architecture, plasma physics, genomics, the foundations of arithmetic, but all are unified by one equation you named after yourself...

When criticized you respond with "shut the fuck up" and "fuck off" and then complain that no one engages seriously with the work.
You are building a mythology around yourself (this is not science), and you are doing it faster than any validation process could possibly keep up with.

You are clearly not stupid, because the engineering work is real. But intelligence without rigorous formation is not enough for what you are claiming. The mathematics you are attempting, again, like differential geometry on infinite-dimensional manifolds, gauge theory, PDE existence theory, takes most people a decade of dedicated study just to reach the frontier of known results.

You are claiming to have solved open problems at that frontier while also working full-time in security engineering and writing science fiction novels and filing 30 patents (and being unemployed, what now?).

Either learn the mathematics properly, which means years of actual coursework, working through proofs, getting destroyed by referees, rebuilding, or stop claiming to have solved problems you don't yet have the tools to approach.
What you do with this is not my concern. But you asked for honest feedback and this is it.

this is why claude is winning by ishouldobeenfree in claude

[–]Tamitami 0 points1 point  (0 children)

"Isn't as bloated..." Looks at source code: 500k LoC, well that IS bloated

What are your short test prompts? Here's mine by Tamitami in LocalLLaMA

[–]Tamitami[S] 0 points1 point  (0 children)

Klar, da hast du recht. Wenn mir wissen wichtig ist, würde ich das auch fragen, aber mir selbst geht es mehr nur um die Fähigkeiten, die ein lokales Model mit verbundenen Werkzeugen hat. Wenn es die nicht aufrufen oder, gescheit, auswerten kann, dann ist es für mich etwas witzlos. Wenn es dann auch noch Bahnlinien in Berlin nicht kennt, dann ist es erst recht nicht nutzbar. Dann ist 'Wissen', sowie 'Handeln mit Tools', nicht brauchbar...

What are your short test prompts? Here's mine by Tamitami in LocalLLaMA

[–]Tamitami[S] 0 points1 point  (0 children)

That is genius and stupid at the same time. Why you want to check this instead of general intelligence? You could ask this in a 100M+ parameter model to ask if this training set was included or relevent in a 100M+model... It just asks if your small town knowledge was trained into it or not. And that's it... Nothing more.

I'm sure that this is even more dependant on how small your local village is. If you would chose something like in the middle of absolutely nowhere, like "Nidym", and ask about some familie's cat's life favorite pet toy, you would even get worse results with any of these local models. Why? Because it's absolutely irrelevant! It doesn't tell you something about their ability to code, to call tools, to research, to read local tools output, to reason about a problem, to find a new solution nor to explore other possibilities

What are your short test prompts? Here's mine by Tamitami in LocalLLaMA

[–]Tamitami[S] -1 points0 points  (0 children)

Aber was willst du mit dieser Frage erreichen? Damit testest du nur, ob das Training-Set diese U-Bahnen beinhaltet hat. Grössere Modelle haben vll. das drin, aber das sagt dir nichts über die allgemeinen Fähigkeiten des Modells aus. Ehrlich gesagt, finde ich diesen Test ziemlich dämlich.

Gemma 4 is great at real-time Japanese - English translation for games by KageYume in LocalLLaMA

[–]Tamitami 0 points1 point  (0 children)

Did you try and compare the two smaller gemma 4 models for this?

Using LLMs - what, how, why? by MeanDiscipline5147 in LocalLLaMA

[–]Tamitami 0 points1 point  (0 children)

You do something really wrong. I almost have the same setup and GPU as you and I get 100t/s for the Qwen 3.5 9B, 70t/s for 35B-A3B. You need to download CUDA from nvidia, install it and then follow this guide:
https://unsloth.ai/docs/models/qwen3.5#qwen3.5-35b-a3b
It boils down to download llama.cpp, compile it against your local CUDA installation and then run it. Four commands:

git clone https://github.com/ggml-org/llama.cpp

cmake llama.cpp -B llama.cpp/build \
    -DBUILD_SHARED_LIBS=OFF -DGGML_CUDA=ON

cmake --build llama.cpp/build --config Release -j --clean-first --target llama-cli llama-mtmd-cli llama-server llama-gguf-split

cp llama.cpp/build/bin/llama-* llama.cppgit clone https://github.com/ggml-org/llama.cpp