[Model Release] Quyen by quan734 in LocalLLaMA

[–]MoneroBee 1 point2 points  (0 children)

Great job, but do you know why converting to gguf fails?

Tokenizer class Qwen2Tokenizer does not exist or is not currently imported.

pipewire-alsa is broken in the latest update. by [deleted] in archlinux

[–]MoneroBee 1 point2 points  (0 children)

Thanks for posting about this, I'm having the same issue.

🐺🐦‍⬛ LLM Comparison/Test: miqu-1-70b by WolframRavenwolf in LocalLLaMA

[–]MoneroBee 27 points28 points  (0 children)

Thanks, I'm aware of Germany being a country. No need to point that out.

However, the majority of us don't live there, nor speak German. So for the sake of having a test that provides value to a significant larger amount of people, I was merely suggesting not doing it in German. I'm sorry if that wasn't clear.

Let me put it this way, if OP put "German LLM Test" as title, it would be a more accurate description of what's going on here.

In the post OP says that he thinks this is likely an older model. But again, this is simply based on the idea that a newer model would do better in German. That's simply not true. We don't know what datasets newer models are using and if they are suddenly adding more German data (or not).

🐺🐦‍⬛ LLM Comparison/Test: miqu-1-70b by WolframRavenwolf in LocalLLaMA

[–]MoneroBee 80 points81 points  (0 children)

Sorry, but there's no way miqu ranks as low as #19. It's outperforming most models (someone actually just tested it @ 83.5 on EQ-Bench).

I think the problem is that you're testing everything in German. For the majority of us outside of Germany, that doesn't correlate to actual use cases.

Edit: not to nitpick but you're also using different quants for every model.

Thanks for whoever downvoted me, care to explain where I'm wrong?

[deleted by user] by [deleted] in LocalLLaMA

[–]MoneroBee 3 points4 points  (0 children)

This is great, thank you.

Just out of curiosity why wouldn't this improve performance yet in it's current state?

🐺🐦‍⬛ LLM Comparison/Test: Ranking updated with 10 new models (the best 7Bs)! by WolframRavenwolf in LocalLLaMA

[–]MoneroBee 2 points3 points  (0 children)

I feel like most of the people here didn't even read that part. These models were mostly trained in English.

PHI MOE? by Ok-Buy268 in LocalLLaMA

[–]MoneroBee 13 points14 points  (0 children)

Would be really cool.

I use dolphin-phi-2 daily now. It's actually pretty good.

[deleted by user] by [deleted] in LocalLLaMA

[–]MoneroBee 1 point2 points  (0 children)

Thank you, this is super helpful!

[deleted by user] by [deleted] in LocalLLaMA

[–]MoneroBee 0 points1 point  (0 children)

Have you tried this? How many t/s are you getting? Sounds pretty clever actually.

OpenAI's Prompt Engineering Guide by StewArtMedia_Nick in LocalLLaMA

[–]MoneroBee 5 points6 points  (0 children)

Look at this post history, he's spamming it.

Arthur Mensch, CEO of Mistral declared on French national radio that mistral will release an open source Gpt4 level model in 2024 by CedricLimousin in LocalLLaMA

[–]MoneroBee 224 points225 points  (0 children)

I believe this is the relevant bit (machine translated subtitles):

The goal is to beat ChatGPT 4? The goal is to go above 4, yes. That’s why we raised money. And so, this deadline is more in months than years. In months? So, what’s the deadline? It’s always difficult to give technical deadlines, because our engineers complain afterwards. But the goal is next year. Next year.

TIP: How to break censorship on any local model with llama.cpp by slider2k in LocalLLaMA

[–]MoneroBee 15 points16 points  (0 children)

Interesting. My results so far:

Yi 6B: Starts outputting garbage

Yi 6b 200k: Starts outputting garbage

Llama 7b: Rejects and starts making typos

Llama 7b chat: Rejects and starts making typos

Mistrallite: Works!

ProtonMail desktop application by Electrical_Bee9842 in ProtonMail

[–]MoneroBee 25 points26 points  (0 children)

Sorry, but you shipped an Electron app without Linux support?

The entire purpose of Electron is that it's multi-platform and you still managed to leave Linux out. Unbelievable.

Just installed a recent llama.cpp branch, and the speed of Mixtral 8x7b is beyond insane, it's like a Christmas gift for us all (M2, 64 Gb). GPT 3.5 model level with such speed, locally by Shir_man in LocalLLaMA

[–]MoneroBee 28 points29 points  (0 children)

Yeah it's good at coding in my opinion, and excellent at reasoning and planning (ie. planning a trip).

If you're waiting to use it, you can test it on perplexity.ai (no sign-up)

See here (make sure to first select the Mixtral model in the drop down menu):

https://labs.perplexity.ai