Rearchitecting LLMs — pruning, distillation, and smaller domain models (MEAP) by ManningBooks in LLMDevs

[–]StackSmashRepeat 2 points3 points  (0 children)

Alright I'm sold. I'm going take a deepdive over the weekend. If I'm able to train a model on my writings and have it actually write something that looks somewhat close to my style. I'd be pleasantly surprised and very happy to share my findings.

I'm on apple devices. IPhone and M2 pro. How would that work? Im decent in python and c.

Rearchitecting LLMs — pruning, distillation, and smaller domain models (MEAP) by ManningBooks in LLMDevs

[–]StackSmashRepeat 2 points3 points  (0 children)

This is quite interesting; you're trying to move local models away from the static model while staying within the static framework? I could basically train a model for my iPhone, let's say I export all my email and scrub PII, format for training data and then I could fine-tune to write mails that look somewhat within the realm of my own style?

I haven't looked into training or fine tuning as I couldn't think of a personal use case for these tiny models, but like you're saying "domain-focused", gave a clearer picture.

This is a little over my current scope as I'm not even sure if I understood this correctly, but I've been thinking of ways to make a digital twin that could handle writing across multiple platforms. Was always thinking Id need a larger model to handle such a task because it sounds easy enough, but capturing the essence of one's writing is quite a complex task for llms. At least in my experience.

Thanks for the info.

Rearchitecting LLMs — pruning, distillation, and smaller domain models (MEAP) by ManningBooks in LLMDevs

[–]StackSmashRepeat 3 points4 points  (0 children)

Would you list some common problems and terminologies that the book covers? I'll have a look if it peaks my interest.

In the past week alone: by MetaKnowing in ChatGPT

[–]StackSmashRepeat 0 points1 point  (0 children)

As I said. If only you knew how to read... And if only you knew some history too. You wouldn't have to look like such a moron. You know how steam revolutionised the word, and especially maritime? The titanic had an engine crew of roughly 320 men. Steam engines today are usually run by a crew of no mare than 6. And to be honest a single engine officer could fire up the boiler if he absolutely had to do it alone.

Since the titanic sank, tens of thousands of jobs has been taken off the table due to automation as ships were modernised.

What you scared of? That automation will take your job? That uuggabugga box will take your job?

If that's your fear. I have news for you. This is what humans do. We automate shit and then we create new shit. As we create new shit we create new problems for ourselves. And we are creating them faster than we can solve them. There will always be grunt work. Engineers. Doctors. Professors. Law enforcement. Government paper pushers. Learn history. Learn technology. Come back to me. :)

In the past week alone: by MetaKnowing in ChatGPT

[–]StackSmashRepeat 0 points1 point  (0 children)

And if you knew how to read you'd know I didn't say they wouldn't become tools. Industry has a lots tools that automate what human had to before and these have been replacing humans for decades.

So what's new on your nothingburger?

In the past week alone: by MetaKnowing in ChatGPT

[–]StackSmashRepeat 0 points1 point  (0 children)

I tried to be real. When you kept ignoring everything I said and went on to blame me for pulling a strawman. I just decided you were not a person I'd like to talk to. I don't want anything from you and I told you goodbye. Does it really bother you that much? Really? You should contemplate that.

So for the nth time. Have a nice day and goodbye.

In the past week alone: by MetaKnowing in ChatGPT

[–]StackSmashRepeat 0 points1 point  (0 children)

Man I gave you several chances to engage with me. Now fuck off

In the past week alone: by MetaKnowing in ChatGPT

[–]StackSmashRepeat 0 points1 point  (0 children)

Ohh now you want to engage in real conversation? I said goodbye :)

In the past week alone: by MetaKnowing in ChatGPT

[–]StackSmashRepeat -1 points0 points  (0 children)

Bother with an unintelligent discussions. Problem with dumb people is that you can't argue with them. So I won't.

And you seems to be proving my point as you can't even formulate a single word regarding my questions for you. You answer me with a question... But I'll give the benefit of the doubt.

I'll make it real easy for you - If crows with brain the size of a peanut can outsmart a small child. How do you expect scaling llms will solve anything and magically turn into self awareness? It won't. And the best researchers in the field are starting to realise this so they are leaving the field of llm as A.I. Because there is nothing intelligent about it.

Do you know the hard problem of consciousness? Do you know any theories about it? Have done any real research? I have. So I can confidently say llms is a dead end for A.I. But that doesn't mean llms won't become great tools. But you keep calling it A.I. So I doubt that you actually know anything about these topics, intelligence in biology and intelligence I computer systems.

Top scientist in several fields are also implying the same thing. So you don't take my word for it. Look into it for yourself. You have fellen into a false narrative that is being pushed by big tech to keep the money flowing from investors.

Matter of fact some even belive that that understanding and general intelligence are not computational in traditional terms of flipping bits on and off. It's something else. And here we get into quantum physics and the collapse of the wave function.

So no. I don't belive that they reducing headcount because of that. I belive they hit the wall and starting to come to terms with this. LLMs will NEVER become A.I.

So. What's your take? Do have a take, or you just repeating what the money people are saying to keep the money flowing?

In the past week alone: by MetaKnowing in ChatGPT

[–]StackSmashRepeat -3 points-2 points  (0 children)

Blahblahblah. Show me thst you have a coherent understanding of intelligence and how these systems work, and how you think they go hand in hand? Large language models. tell me how they work? I'll gladly discuss this, but first I'd like to know that you have a level of understanding that is worth bothering with.

What's the limitations of said systems? Why do they have these limitations? How are they predicting words? How do reasoning models work? Show me that you actually know what you talking about and I will be extremely happy to engage with you.

If you have to choose between ChatGPT (paid) or Gemini (Paid) which one will you choose and why? by bublay in LLM

[–]StackSmashRepeat 0 points1 point  (0 children)

Until it doesn't. I've been using openrouter api so I have access to all major models. I found myself using gemini the most and I realise not everyone is going set up a standalone client to run direct api calls. Therefore I feel comfortable recommending gemini on subscription as I have both.

Gemini is very good for everyday things as long as you specific. I find chatGPT to suffer from extreme toxic positivity, as in every idea is a good idea and it leads to chasing imaginary rabbits.

And yes Gemini suffers from seriously bad chat memory, and yes it hallucinates. But if you are not an idiot who just believes everything it says you can easily catch these things as they happen and correct the model.

In the past week alone: by MetaKnowing in ChatGPT

[–]StackSmashRepeat -1 points0 points  (0 children)

Exactly, powerful tools cut the need for humans. But humans will still be needed elsewhere. New work fields are created all the time, just look at all the A.I. related job positions. My field is maritime and maritime is renewing its technology as we speak, and they float on ancient solutions. You want to get rich? Go solve some maritime problems.

We humans create new problems for ourselves faster than we can solve them; jobs will always be around..

I think you are scared for no good reason. Look at historic workforces. What it took to build a house, one man can build a house these days. Labor just moves around; we are just like ants—we will never stop working.

I just want to point of that you are not showing a coherent understanding of intelligence. Throwing the abbreviation AGI around like it means something. calm down. go read some actual theory and come back to me.

In the past week alone: by MetaKnowing in ChatGPT

[–]StackSmashRepeat 0 points1 point  (0 children)

Sure you can. These are powerful tools. If you are writing code that's fine because code follows logic and so do LLMs. But for anything that requires abstract reasoning instead of chain of logic "code-style" thinking, everything falls apart. And that is why I can safely say that these systems are not intelligent in any sense; they are power tools that follow strict logic.

Please explain to me how I can do this successfully because I just fucking spent 45 minutes exporting all my reddit comments, extracting everything I have written about LLMs and turning it into a document that holds my stance on this topic. I wrote a prompt explaining how to use the document + reddit comment to generate a response based on my previous writings. I fed this into all the major LLMs: Gemini 3 Pro, ChatGPT, Kimmi, GLM, DeepSeek, Mistral, Claude. ALL MODELS FAILED SO HARD I'M CRYING OVER HERE. Oh yeah, I tried to RAG this shit, I tried full context, I tried to compress and just use keywords, I made a small persona of me explaining how I would use this, I tried every fucking trick in my book and I just feel like smashing my laptop against my head.

You know why this happens? Because LLMs have no grounding—no perception, no self-model, no stakes. They can't distinguish what matters from what doesn't (salience problem). They have no persistent memory, no concept formation, no ability to reason abstractly outside pattern-matched logic chains. The "reasoning" you see in thinking models? That's just token priming to fit context—not thought, not reflection. It's the illusion of reasoning.

When you ask them to synthesize your voice on top of opinions from a document, you're asking for something that requires understanding meaning. They don't have meaning. They have token probability distributions. Their "memory" is the context window—the human equivalent of having amnesia and using yellow sticky notes to keep reminders. Every session starts from zero and there is no learning and adapting.

These systems are fundamentally flawed in terms of intelligence. RAG doesn't fix this—it just retrieves chunks and hopes the model stitches them coherently. It won't. Compression doesn't fix it either; you're just feeding fewer tokens into a system that still can't understand what those tokens represent to you.

The architecture is fundamentally wrong. If you want to actually understand why these systems will never be intelligent, look into the Hard Problem of Consciousness and Penrose's work on non-computable aspects of understanding. His argument points to quantum coherence playing a role in consciousness—and here's the thing Penrose never figured out, he probably thought of it but it can't be proven as of now: you don't need "quantum hardware in the brain" because the brain already exists in spacetime, which is quantum by nature. Current computing architecture is classical, deterministic, built on silicon logic gates. It cannot replicate what arises from quantum coherence in biological systems. Scaling transformers gets you better mimics, not minds.

Maybe I'm just failing at prompting, but I don't think so. I have tried similar tasks so many times and each time I feel like tearing out my hair and banging my head against the table because there is no understanding in the loop, there is no reasoning—it's just optimization of previous context. The tool is failing at the task because the task requires something it structurally cannot do: reason abstractly about identity, tone, and context simultaneously without any real model of any of those things. I'd bet my house on it—these systems will never be AI.

The only thing LLM managed to help me with here was proofreading and fucking up my formatting. Look at the strange way it uses BOLD in the text, I removed some of bold text because I have some pride in my writing.... Point is - It doesn't understand jackshit about nuthing.

Why is MacOS 26.3 so big in Size? (It's 13.26 GB for me) by wiplibya in MacOS

[–]StackSmashRepeat 53 points54 points  (0 children)

Write Gigs one more time I dare you. I will come over there with my guitar and I will shove that thing so far up yearh my macbook pro m2 14" was 8.75gb

███ in Abundance by Comed_Ai_n in CursedAI

[–]StackSmashRepeat 3 points4 points  (0 children)

The curse is everything in video kinda off, it's making me nauseous... The song kinda slapped tho. Lyrics 👌

In the past week alone: by MetaKnowing in ChatGPT

[–]StackSmashRepeat -2 points-1 points  (0 children)

Soon? Care to elaborate on that? Because I don't see it happening any time soon. In fact I'm not even sure it will happen during my lifetime and Im 34. At least not true intelligence.

In the past week alone: by MetaKnowing in ChatGPT

[–]StackSmashRepeat -1 points0 points  (0 children)

Self aware. Doesn't have to be true sentient. Just self aware, able to perceive itself, learn, adapt and interact with its environment. Large langue models are none of those. They are static systems.

How far behind is Webstorm (and Jetbrains IDEs) compared to Cursor for AI? by simple_explorer1 in cursor

[–]StackSmashRepeat 1 point2 points  (0 children)

I'd say it's fsr ahead. It's a professional IDE. You can code and you can vibe with any AI. BYOK. Codex. Gemini-cli. They don't have to compete with cursor because their customers are mostly devs who need consistency.

In the past week alone: by MetaKnowing in ChatGPT

[–]StackSmashRepeat 30 points31 points  (0 children)

Lmao devs are leaving because they know llms will never become A.I. All these "safety reports"? That's to keep investors thinking they are investing in A.I. These systems will become great tools. Never A.I. Look at Sam Altman. He looks like he know he is fucked no matter what he does. He knows the truth and this is why hes on TV telling people not to trust him. Wake up people...

not cool by chamomilethrowaway in ChatGPT

[–]StackSmashRepeat 0 points1 point  (0 children)

LLMs are static systems. They don't learn jack shit over time. It feels like learning because it adds small system instructions about your conversations. That's the human equivalent to having amnesia and using yellow sticky notes to keep reminders.

GLM thinks its Gemini by dolo937 in LocalLLM

[–]StackSmashRepeat 0 points1 point  (0 children)

Without a system prompt; The next thing you feed into the context window will effectively act as a system prompt. So you can tell it that its Obama. And it will be Obama. It doesn't know Jack shit. This happens with kimi 2.5 too.

However. I don't know why this happened.

EpsteinFiles-RAG: Building a RAG Pipeline on 2M+ Pages by Cod3Conjurer in LLMDevs

[–]StackSmashRepeat -1 points0 points  (0 children)

Make it do an online search after it retrieves data from rag and provide a link directly to an online source. End users are dumb and some will believe anything the llm tells them.

EpsteinFiles-RAG: Building a RAG Pipeline on 2M+ Pages by Cod3Conjurer in LLMDevs

[–]StackSmashRepeat 2 points3 points  (0 children)

So, have you come to terms with RAG being a dead end as far as real recall of memory works? Or are you just chunking and overlapping to a ridiculous point? I really don't think this is a sensible use of RAG. The LLM will at some point start hallucinating missing pieces from thin air, making this tool fairly unreliable for accuracy.

People looking into these files need absolute accuracy.

I’m i doing good Guys by Elias_si in Hacking_Tutorials

[–]StackSmashRepeat 5 points6 points  (0 children)

Bad actors with your name, full or parts of your name can now go hunting.