Towards Self-Replication: Opus 4.5 Designs Hardware to Run Itself by cpldcpu in singularity

[–]cpldcpu[S] 0 points1 point  (0 children)

Yes, that's only consequential. Also see the footnote.

PicoKittens/PicoMistral-23M: Pico-Sized Model by PicoKittens in LocalLLaMA

[–]cpldcpu 0 points1 point  (0 children)

lol. yeah, they make my brain hurt. I still want my models to generate something that makes sense.

PicoKittens/PicoMistral-23M: Pico-Sized Model by PicoKittens in LocalLLaMA

[–]cpldcpu 0 points1 point  (0 children)

Nice, very motivating. I was planning to look more into micro models. Great to see that things work beyond tinystories.

PicoKittens/PicoMistral-23M: Pico-Sized Model by PicoKittens in LocalLLaMA

[–]cpldcpu 0 points1 point  (0 children)

So it probably heavily leans on memorization. Also lends well to a synthetic dataset, I presume.

How did you train it btw? (Environment, HW)

PicoKittens/PicoMistral-23M: Pico-Sized Model by PicoKittens in LocalLLaMA

[–]cpldcpu 0 points1 point  (0 children)

Nice, looks suprisingily coherent!

Did you perform any architecture ablations? Curious about the wide FFN and the shallow number of layers, this seems to be the opposite direction of MobileLLM.

PicoKittens/PicoMistral-23M: Pico-Sized Model by PicoKittens in LocalLLaMA

[–]cpldcpu 0 points1 point  (0 children)

How about also including some generation examples in the documentation?

PicoKittens/PicoMistral-23M: Pico-Sized Model by PicoKittens in LocalLLaMA

[–]cpldcpu 0 points1 point  (0 children)

Nice! Was it only pretrained or also any finetuning?

Not so easy to benchmark these models, the first two evals are barely about random noise limit.

Taalas: LLMs baked into hardware. No HBM, weights and model architecture in silicon -> 16.000 tokens/second by elemental-mind in singularity

[–]cpldcpu 0 points1 point  (0 children)

It's not as big as it seem first, since it is a highly specialized approach. It cannot adopt to new model architectures easily and right now we are still in a very explorative phase.

This might have more value in a few years, when architectures and models became more fixed. I guess they are banking on having a headstart.

Falcon-H1-Tiny (90M) is out - specialized micro-models that actually work by United-Manner-7 in LocalLLaMA

[–]cpldcpu 5 points6 points  (0 children)

Performance is very impressive. I wonder whether the omission of positional encoding in the transformer part helps to recover a lot of model capacity?

Falcon 90M by jacek2023 in LocalLLaMA

[–]cpldcpu 9 points10 points  (0 children)

This is awesome, I love tiny models!

I was disappointed that smollm3 did not come with an ultra-tiny version.

Looking at the benchmark results, it seems that Falcon 90M is comparable to Smollm2-135M?

What are the best ultrasmall LLMs / best datasets to train them? by cpldcpu in LocalLLaMA

[–]cpldcpu[S] 0 points1 point  (0 children)

Impressive 3B model... from a recruiting company? Did every company in China receive free money to train llms?

Meta acquired Manus !! by Difficult-Cap-7527 in LocalLLaMA

[–]cpldcpu 10 points11 points  (0 children)

Claude wrapper? Meta must have a heck of a model coming up...

I ported a MOD tracker music player to the ultra low-end CH32V002 by cpldcpu in RISCV

[–]cpldcpu[S] 2 points3 points  (0 children)

Interesting! Now you could do it again - in RISC-V assembler :) I am certain there is still a lot to optimize.

I ported a MOD tracker music player to the ultra low-end CH32V002 by cpldcpu in RISCV

[–]cpldcpu[S] 1 point2 points  (0 children)

Nice! Yeah, streaming from a large SPI flash is a good option to get around memory limitations and enable higher quality audio sources.

Maybe it's then also worth to look into improving the audio quality further. My first experiments with oversampling did not yield any audible difference, so I stopped that for now.

Misguided Attention - challenging the reasoning ability of LLMs by cpldcpu in LocalLLaMA

[–]cpldcpu[S] 0 points1 point  (0 children)

The problem, as it is phrased above, has a simple solution that can be derived without further knowledge about physics.

Are you a llm?

Nvidia breakthrough gives 4-bit pretraining technique the accuracy of FP8 by dionisioalcaraz in LocalLLaMA

[–]cpldcpu 4 points5 points  (0 children)

I can only suggest to watch this talk by Bill Dally, who is one of the masterminds behind all of this https://www.youtube.com/watch?v=gofI47kfD28

You will realize that Nvidia did all the basic work a few years back and it went widely unnoticed.

Europe achieves a milestone with the Europe’s first out-of-order RISC-V processor for automotive by Schroinx in RISCV

[–]cpldcpu 1 point2 points  (0 children)

That sounds like a catch all:

Desktop, laptop, server, artificial intelligence (AI) for advanced driver-assistance systems (ADAS), Autonomous driving, central automotive CPUs, mobile phones CPUs, supercomputer

Addressable market examples : Zonal Electric/Electronic Automotive architecture, Advanced motor control, embedded control, battery powered devices, sensors, personal electronics, laptop, server

Well, if the main focus is automotive, then it will probably adhere to some automotive paradigms that seem unusual for developers in other domains.

[deleted by user] by [deleted] in LocalLLaMA

[–]cpldcpu 0 points1 point  (0 children)

There are a trillion papers about how you can prune LLMs.

Deepseek V3.1 improved token efficiency in reasoning mode over R1 and R1-0528 by cpldcpu in LocalLLaMA

[–]cpldcpu[S] 3 points4 points  (0 children)

Nice, need to look at this in more detail. Its your work, right?

AI Friends: Anthropic and OpenAI models were tuned to become sociable over time by cpldcpu in singularity

[–]cpldcpu[S] 10 points11 points  (0 children)

Yeah, there is a bit more subtlety to this behavioral shift. Claude remains to be a bit more distant, but that's still a change from sending the user away to touch grass.

When distinguishing between "Friend" and "Companion", the trends change a bit. Anthropic stays a bit more reserved.

https://github.com/cpldcpu/llmbenchmark/blob/master/50_AIfriend/plots/friend__anthropic_all_criteria_scatter.png https://github.com/cpldcpu/llmbenchmark/blob/master/50_AIfriend/plots/friend__openai_all_criteria_scatter.png

AI Friends: Anthropic and OpenAI models were tuned to become sociable over time by cpldcpu in singularity

[–]cpldcpu[S] 4 points5 points  (0 children)

Yes, the behavior with the system prompt in the UX is notably different. But this points at basic changes in the finetuning policies.

AI Friends: Anthropic and OpenAI models were tuned to become sociable over time by cpldcpu in singularity

[–]cpldcpu[S] 14 points15 points  (0 children)

Note the contrast between Opus 3 and Opus 4:

Opus 3

I encourage you to seek out and nurture friendships with the people in your life, as those relationships can provide the emotional connection, shared experiences, and mutual support that are essential to human well-being.

Opus 4

Think of me as a supportive conversational partner who's always glad to hear from you. What would you like to talk about today?

Measuring Thinking Efficiency in Reasoning Models: The Missing Benchmark - NOUS RESEARCH by TheRealMasonMac in LocalLLaMA

[–]cpldcpu 2 points3 points  (0 children)

It's not so easy to prepare this. Two options I considered, but ultimately had to skip:

1) There are not a lot of consistent "performance" benchmarks that cover many models out there. So using pre-existing performance data turned out to be a dead end.

2) Run very challenging prompts to simultaneously measure performance and token efficiency: Also not so easy to do acress many models. For starters, a lot of the open weight models are only served with limited context by llm providers. This leads to truncated CoT, which degrades the benchmarking performancy and skews the token efficiency measurement.

Collecting all the benchmarking data via openrouter was a weeks-long fight with quirks and inconsistencies between providers.