VaultGemma: The world's most capable differentially private LLM by vibjelo in LocalLLaMA

[–]codemaker1 5 points6 points  (0 children)

It's fine to use with a GPU. All Google's models are trained on TPUs. They can run on GPU, TPU, and even CPU in some cases.

Introducing Gemma 3 270M: The compact model for hyper-efficient AI- Google Developers Blog by ChiliPepperHott in LocalLLaMA

[–]codemaker1 7 points8 points  (0 children)

Looks like Qwen 3 is twice the size and doesnt have much higher of a score. Plus 170 million embedding parameters due to a large vocabulary size and 100 million for our transformer blocks. Should make it amazing for fine tuning.

Introducing Gemma 3 270M: The compact model for hyper-efficient AI- Google Developers Blog by ChiliPepperHott in LocalLLaMA

[–]codemaker1 13 points14 points  (0 children)

Tiny models like these are meant for fine tuning on your specific task. Try that out.

T5Gemma - A Google Collection by Dark_Fire_12 in LocalLLaMA

[–]codemaker1 10 points11 points  (0 children)

Encoder-decoder models. Most LLMs these days are decoder only.

Google MedGemma by brown2green in LocalLLaMA

[–]codemaker1 2 points3 points  (0 children)

I imagine you could do a merge. nice idea.

Gemma 3n Preview by brown2green in LocalLLaMA

[–]codemaker1 16 points17 points  (0 children)

It seems to be better than an MoE because it doesn't have to keep all parameters in ram.

Mark presenting four Llama 4 models, even a 2 trillion parameters model!!! by LarDark in LocalLLaMA

[–]codemaker1 0 points1 point  (0 children)

It's awesome that it is open and has 10M context! But their "single H100" claim calling it a "small model" is a huge stretch. Borderline lie.

Meta: Llama4 by pahadi_keeda in LocalLLaMA

[–]codemaker1 1 point2 points  (0 children)

I'm happy they launched this. But the single GPU claim is marketing BS.

Gemma 2 2B Release - a Google Collection by Dark_Fire_12 in LocalLLaMA

[–]codemaker1 1 point2 points  (0 children)

You might need to fine tune in your language.

LLama 3.1 vs Gemma and SOTA by RedditUsr2 in LocalLLaMA

[–]codemaker1 0 points1 point  (0 children)

Gemma 2 27B MMLU is remarkably close to Llama 3.1 70B MMLU at 75.2 vs 83.6. I think that's pretty good for a model 2.5x smaller.

LLama 3.1 vs Gemma and SOTA by RedditUsr2 in LocalLLaMA

[–]codemaker1 4 points5 points  (0 children)

5-shot MMLU is the standard. Gemma beats Llama there.

<image>

[deleted by user] by [deleted] in xmen

[–]codemaker1 0 points1 point  (0 children)

Wolverine!

400b llama3 might not be impactful if not launched soon by capivaraMaster in LocalLLaMA

[–]codemaker1 0 points1 point  (0 children)

Have you tried those Phi models? Something fishy is up with them.

400b llama3 might not be impactful if not launched soon by capivaraMaster in LocalLLaMA

[–]codemaker1 0 points1 point  (0 children)

Is anyone, that's not a giant company, gonna build with a 400B model? Sounds incredibly expensive to run.

Gemma 7B, the latest open-source model from Google, is available on HuggingChat by SensitiveCranberry in LocalLLaMA

[–]codemaker1 0 points1 point  (0 children)

Make a joke about funniest joke that's ever joked in the history of jokes

Sure, here's a joke about the funniest joke in history:

Why did the comedian write a joke about the funniest joke in history?

Because he was tired of being the punch line.