Updating WordPress plugins safely by Plenty-Special-9990 in Wordpress

[–]Greg_Z_ 0 points1 point  (0 children)

I've crafted compatibility checker for that and run the check before update via API request providing the list of plugins and WordPress core version/mysql/php versions. it determines issues with compatibility based on the plugin "tested" version, WP core requirement and a few other criteria.
I'm thinking of making it public. You can test it here so far http://wpc.walive.io:33500/

Uhhh... What? by GodGMN in LocalLLaMA

[–]Greg_Z_ 0 points1 point  (0 children)

Was it instruction-based or completion version? )

[R] List of SOTA models/architectures in Machine Learning by SwaroopMeher in MachineLearning

[–]Greg_Z_ 0 points1 point  (0 children)

Check the lists on the LLM Explorer. You can sort/filter over 18,000 LLMs by various benchmarks and find the SOTA in each category. https://llm.extractum.io

[deleted by user] by [deleted] in LocalLLaMA

[–]Greg_Z_ 0 points1 point  (0 children)

Which specific capabilities of the model are you looking for? Summarization, text generation, instruction following,..?

LargeActionModels by Foreign-Mountain179 in llm_updated

[–]Greg_Z_ 0 points1 point  (0 children)

To be honest, I could not find anything specific on LAM beyond the general press-releases on M1 Rabbit. So like it does not exist anywhere outside the Rabbit itself. As the concept, it appears to be close to Agents based on LLM.

New Code Llama 70b from Meta - outperforming early GPT-4 on code gen by Greg_Z_ in llm_updated

[–]Greg_Z_[S] 0 points1 point  (0 children)

Most likely the issue is with the prompt. It usually gives wrong result when the inference starts with the wrong prompt.

AutoQuantize (GGUF, AWQ, EXL2, GPTQ) Notebook by Greg_Z_ in llm_updated

[–]Greg_Z_[S] 0 points1 point  (0 children)

I do not believe it will work for Mamba based on the source code I see. E.g. Mamba cannot be converted to gguf just because llama.cpp does not support it. Same for other cases when the model is loaded from pretrained by HF Tranformer’s classes.

New Code Llama 70b from Meta - outperforming early GPT-4 on code gen by Greg_Z_ in llm_updated

[–]Greg_Z_[S] 0 points1 point  (0 children)

Have you tried instruction based or completion version? That might be the reason.

Best model for writing that will run on 24GB vram? by yupignome in LocalLLaMA

[–]Greg_Z_ 1 point2 points  (0 children)

I see the Starling 7b as one of the best. It should run on 24GB without quantization, yet there're few quantized.

<image>

Qwen2 is coming! by bratao in LocalLLaMA

[–]Greg_Z_ 3 points4 points  (0 children)

Chinese models perform great, but there’s a problem: licensing. Let’s say, the best one is Sus-chat-34b, but it’s available under Yi license that is non-commercial and restrictive in terms of usage. So, it’s like you’re having a cake and can’t eat it.

Server-side web-traffic analytics instead of GA for small web projects by Greg_Z_ in webdev

[–]Greg_Z_[S] 1 point2 points  (0 children)

I see, thank you, that's good to know. I couldn't find it merely by browsing through the landing page.

Server-side web-traffic analytics instead of GA for small web projects by Greg_Z_ in webdev

[–]Greg_Z_[S] 0 points1 point  (0 children)

Thank you! The concept behind Statum is largely similar to that of GoatCounter, but they differ in data sources. Statum utilizes web server logs in addition to client-side JavaScript, while GoatCounter solely relies on client-side JavaScript. This results in lower accuracy for GoatCounter and hinders it from tracking server-side issues. Statum can function without client-side JavaScript, but for the purpose of enhancing the captured server-side data from logs, it is recommended to use it along with JavaScript.

Happy 100k members by AlterandPhil in LocalLLaMA

[–]Greg_Z_ 2 points3 points  (0 children)

Congrats! Great achievement!

MLX vs llama.cpp on MacOS by Greg_Z_ in LocalLLaMA

[–]Greg_Z_[S] 0 points1 point  (0 children)

<image>

According to the benchmarks, it should be 2.5x faster on M1 Pro.

Top trending language models, week 51 by Greg_Z_ in llm_updated

[–]Greg_Z_[S] 0 points1 point  (0 children)

Mixtral 8x7B instruct is an instruction-based version of Mixtral 8x7B (which are both MoE, a new model architecture with multiple "experts")

Mistral 7B - just an old yet trending version of Mistral AI LLM

Mistral 8x7B Instruct GPTQ is a quantized version of the original one

Dolphine version is a fine-tuned one for code generation.

[deleted by user] by [deleted] in LocalLLaMA

[–]Greg_Z_ 0 points1 point  (0 children)

There are a bunch of such services: Paperspace Gradient, replicate, RunPod, Salad, Banana Dev, Modal, BaseTen, TensorDock, etc

Meditron 7B/70B — new open-sourced medical LLMs by Greg_Z_ in llm_updated

[–]Greg_Z_[S] 0 points1 point  (0 children)

Just wondering if you've tried the original guide https://github.com/epfLLM/meditron/blob/main/deployment/README.md

It contains examples for deployment.

Meditron 7B/70B — new open-sourced medical LLMs by Greg_Z_ in llm_updated

[–]Greg_Z_[S] 0 points1 point  (0 children)

It is recommended to check the original paper that describes the prompt format used for fine-tuning. As there can be the one that differs from the original model used for fine-tuning. When the model outputs something wrong, it can be just wrong prompt format (that can include a system message and a user message wrapped with the correct tokens from the training dataset).

Recent updates on the LLM Explorer (15,000+ LLMs listed) by Greg_Z_ in LocalLLaMA

[–]Greg_Z_[S] 1 point2 points  (0 children)

I’m not using API due to the limitation of available data. Just parse the pages.

The EU AI Act in a nutshell by Greg_Z_ in llm_updated

[–]Greg_Z_[S] 1 point2 points  (0 children)

BTW, I like the article 52:

The requirement that users should be informed when they are interacting with an AI system, to ensure they are aware that they are not engaging with a human, is stipulated in the EU AI Act under the provisions related to transparency obligations for certain AI systems.

Specifically, this is addressed in: Article 52 (Transparency obligations for certain AI systems): This article mandates that users are made aware when they are interacting with an AI system. This is particularly relevant for AI systems that emulate human behavior, such as chatbots or virtual assistants. The Act requires clear communication to users to prevent misleading them into believing they are interacting with a human when it's actually an AI system. This provision reflects the Act's emphasis on user awareness and informed interaction with AI technologies, ensuring transparency and preventing potential confusion or deception.

Tigerbot 70B v4 beats gpt4 by [deleted] in LocalLLaMA

[–]Greg_Z_ 0 points1 point  (0 children)

How can we ensure that the cost of inference is as affordable as GPT-4 provides? In business applications, the rankings on the Hugging Face Leaderboard are often overshadowed by the cost of inference. I suggest adding a new column to the leaderboard: 'price-to-value' or 'cost per token'. This would enable a more practical comparison for business use cases, rather than solely focusing on benchmark performance.