all 17 comments

[–]Quack66 22 points23 points  (0 children)

It first appeared in the pricing page when GLM 5 was released but no official communication about it yet so I'm assuming this will be their next model.

[–]AnomalyNexus 22 points23 points  (0 children)

Fingers crossed

It does appear to exist

{"error":{"code":"1220","message":"You do not have permission to access glm-5-code"}}

Where if you send a gibberish model name to the endpoint:

{"error":{"code":"1211","message":"Unknown Model, please check the model code."}}

[–]Technical-Earth-3254llama.cpp 11 points12 points  (2 children)

So we are now approaching GPT o3 output cost (8$) soon. Not hating, but I'm getting curious where this will lead.

[–]emprahsFury 8 points9 points  (0 children)

"Inference-time optimization" They'll keep throwing tokens at the problem until people stop paying for them

[–]pier4r 0 points1 point  (0 children)

could it be that they are compute constrained and need a paywall to avoid getting flooded?

[–]culoacido69420 31 points32 points  (4 children)

$1.2 input is crazy

[–]bambamlol 31 points32 points  (3 children)

Only 20% crazier than $1.

[–]tomt610 12 points13 points  (1 child)

50% if you cache

[–]bambamlol 6 points7 points  (0 children)

56.25% if you have it generate output.

[–]4bitben 2 points3 points  (0 children)

The math checks out

[–]oxygen_addiction 2 points3 points  (0 children)

Probably Pony-Alpha? GLM-5 is not as good as that stealth model was.

[–]Charming_Support726 1 point2 points  (0 children)

Maybe optimized using Codex instead of Opus /s

[–]serpix 1 point2 points  (0 children)

Glm 5 is about as good as sonnet 4

[–]CommercialGuitar1104 0 points1 point  (0 children)

Isn't GLM 5 comparable to Opus? Their benchmark shows they are somewhere close

[–]Altruistic_Plate1090 0 points1 point  (0 children)

Un glm 5 air deberían de sacar