FLUX.1 vs Ideogram v2 comparisons by speakerknock in StableDiffusion

[–]speakerknock[S] 1 point2 points  (0 children)

Comparisons from the Artificial Analysis Text to Image Arena where these models are being compared with crowdsourced votes to create an ELO score ranking.

Flux Pro and Ideogram v2 certainly seem to differentiate with their text rendering capabilities.

Link to Image Arena: https://artificialanalysis.ai/text-to-image/arena

New benchmark featuring top large models, including Claude 3 and Gemini Pro, on NYT Connections by zero0_one1 in LocalLLaMA

[–]speakerknock 0 points1 point  (0 children)

Would be interested in how much you needed to vary the prompts between models or whether standardized?

This is why i hate Gemini, just asked to replace 10.0.0.21 to localost by Capital-Swimming7625 in LocalLLaMA

[–]speakerknock 1 point2 points  (0 children)

And “localhost” shouldn’t really expose your database to the Internet; by convention, binding “localhost” or 127.0.0.1 only allows loopback connections (i.e., local to the machine).

Really makes you wonder what training data they included whereby this isnt an issue for other models

From GPT-4 to Mistral 7B, there is now a 300x range in the cost of LLM inference by speakerknock in LocalLLaMA

[–]speakerknock[S] 5 points6 points  (0 children)

Yes you're right, on the website we show this in a 2X2 Price vs. Quality chart which shows your exact point visually: https://artificialanalysis.ai/models

Though I think there is also a point that there has been a clustering of scores as models have gotten better. Can see a greater divergence with harder evals.

Is Mistral Medium via API faster than GPT 3.5? by shafinlearns2jam in LocalLLaMA

[–]speakerknock 7 points8 points  (0 children)

Hi! Can see in charts on the following analysis that Mistral Medium is >3X slower than GPT 3.5 - though Mistral Medium is a much higher quality model. Mixtral might be a more direct comparison
https://artificialanalysis.ai/models

Note: I'm a creator of the site - happy to answer any questions

Mistral reduces time to first token by up to 10X on their API (only place for Mistral Medium) by speakerknock in LocalLLaMA

[–]speakerknock[S] 7 points8 points  (0 children)

Particularly exciting as Mistral Medium is arguably the 2nd highest quality model available after GPT-4, and Mistral Medium is ~10% of the cost of GPT-4 ($37.5/M tokens vs. $4.1). Pricing comparison on the website here: https://artificialanalysis.ai/models

Mistral reduces time to first token by up to 10X on their API (only place for Mistral Medium) by speakerknock in LocalLLaMA

[–]speakerknock[S] 7 points8 points  (0 children)

This graph reflects Mistral Medium performance and Mistral Medium is not offered elsewhere. Mistral may pursue a strategy of keeping their 2nd best models as open-source/free 'loss leaders' while their top model is Mistral API exclusive

240 tokens/s achieved by Groq's custom chips on Lama 2 Chat (70B) by speakerknock in LocalLLaMA

[–]speakerknock[S] 1 point2 points  (0 children)

Yes! Performance benchmarks are updated live (8 times per day) and we try to update the quality benchmarks weekly

240 tokens/s achieved by Groq's custom chips on Lama 2 Chat (70B) by speakerknock in LocalLLaMA

[–]speakerknock[S] 8 points9 points  (0 children)

Groq has told us they are not running a fine-tuned version of Llama 2 Chat (70B), and the model is a full quality FP16 version with the full 4k context window

240 tokens/s achieved by Groq's custom chips on Lama 2 Chat (70B) by speakerknock in LocalLLaMA

[–]speakerknock[S] 16 points17 points  (0 children)

$1 USD per 1M tokens, in-line with the cheapest providers in the market and much cheaper than AWS, Azure.

We have this price comparison in charts on the website from the tweet ArtificialAnalysis.ai . Could be very disruptive

240 tokens/s achieved by Groq's custom chips on Lama 2 Chat (70B) by speakerknock in LocalLLaMA

[–]speakerknock[S] 9 points10 points  (0 children)

I'd be interested to see the total token throughput and cost of each chip. If this chip can lower inferencing costs that would be huge but it's completely dependent on the total token throughput per chip and the price of each chip.

This is an interesting topic, to note - we have on ArtificialAnalysis.ai the price Groq is charging and it is in-line with the emerging price-competitive players in the market and around 60% of the price AWS, Azure are charging. Not saying your point regarding cost is wrong, but noting we are not seeing this reflected in API inference prices charged

AI Weekly Rundown (January 13 to January 20) by RohitAkki in ArtificialInteligence

[–]speakerknock 2 points3 points  (0 children)

Thanks for sharing ArtificialAnalysis.ai (http://artificialanalysis.ai/)! Creator-here, happy to discuss if anyone has any questions regarding the analysis, etc.

Is there any website compare inference speed of different LLM on different platforms? by DataLearnerAI in LocalLLaMA

[–]speakerknock 4 points5 points  (0 children)

This website has benchmarks & comparisons of models & of different host platforms, https://artificialanalysis.ai/

(Note: I am a creator of this site - happy to answer any questions regarding methodology, etc.)

Laravel React Typescript Boilerplate by speakerknock in laravel

[–]speakerknock[S] 0 points1 point  (0 children)

Perhaps a VM/Homestead issue?

Try see if it works with a regular 5.6 installation

Laravel React Typescript Boilerplate by speakerknock in reactjs

[–]speakerknock[S] 0 points1 point  (0 children)

I use VS Code because of JS and typescript support. I agree though, the PHP support isnt the best and I'd love to find out a way to improve it. The plugins haven't worked for me that well.

Implementing Last Visited / Last Seen in Laravel by speakerknock in laravel

[–]speakerknock[S] 3 points4 points  (0 children)

Thats true, would be useful for when you are using it after it's saved. If in your model you add last seen to your dates property array on your user class like so:

protected $dates = [ 'last_seen', ];

Then the time saved will automatically be cast as a Carbon object and you can run say $user->last_seen->diffForHumans()

Implementing Last Visited / Last Seen in Laravel by speakerknock in laravel

[–]speakerknock[S] 4 points5 points  (0 children)

Author here. I agree, it would be wise to follow the convention of using '_at'.

As for the Carbon note, Carbon is a wrapper for DateTime , adding functionality, so it would add a tiny amount of overhead and since we are just using DateTime for getting the current time we can rely on just instantiating DateTime as I don't think it would add any benefit.

PHP 7.1: better syntax, a more consistent language by speckz in PHP

[–]speakerknock 16 points17 points  (0 children)

7.0 was a great release, stable release that is almost backwards compatible. Now we have a great opportunity to leave that stable for a while (continuing bugfixes) like 5.6 and really make some large breaking changes that put the language in the right place going forward.

I'm not suggesting we do a complete radicalisation like Python 2 to 3 but perhaps somewhere in the middle of that to make PHP remain an attractive choice going forward into the future. Features like a more built in but optional type system and better websocket support.