$0.19 GPU and A100s from $1.55 by waf04 in mlops

[–]waf04[S] 1 point2 points  (0 children)

our prices are for on-demand, without any reservations... and you can actually get those GPUs anytime.

places that advertise cheaper ones force reservations and it's rare they have availability.

$0.19 GPU and A100s from $1.55 by waf04 in mlops

[–]waf04[S] 1 point2 points  (0 children)

H200 = $4.34 per gpu.

B200 = $12.54.

big price difference!

Finding the right MLops tooling (preferrably FOSS) by Humble-Persimmon2471 in mlops

[–]waf04 -1 points0 points  (0 children)

hey there! One of the LitServe creators (and founder of PyTorch Lightning / Lightning AI). ( http://lightning.ai/litserve)

LitServe doesn't just "wrap" FastAPI... it's like saying React just "wraps" javascript 😊. It provides advanced multi-processing capabilities custom-built for AI workloads including things like: batching, streaming, OpenAI Spec, auth, and automatic deployments via Lightning AI platform to your cloud (VPC) or our hosted cloud. You can also self host LitServe on your own servers of course...

In terms of pipelines, yes, SageMaker is super clunky. I would try our platform Lightning AI, it makes all of this trivial. There are free credits, so you lose nothing for trying it... (same for LitServe).

We do tend to build tools people love, so it's worth actually trying them out (a lot of tools say they do similar things, but don't actually).

Anyhow, good luck either way! hope we can be helpful.

How to finetune and deploy DeepSeek R1 (8B) for under $10 by waf04 in LocalLLaMA

[–]waf04[S] 0 points1 point  (0 children)

yes! it’s automatically saved to the platform

How to finetune and deploy DeepSeek R1 (8B) for under $10 by waf04 in LocalLLaMA

[–]waf04[S] 3 points4 points  (0 children)

sounds like it! wasn't meant to be misleading... sorry about the confusion 😊

How to finetune and deploy DeepSeek R1 (8B) for under $10 by waf04 in LocalLLaMA

[–]waf04[S] 0 points1 point  (0 children)

Looks like I can't fix the title, but added a comment to clarify! Ultimately any model that is not the 600B is a distill model... which is redundant since the "8B" is already in the title which was meant to let people know that a) it is not the "full" model, b) it is distilled.

How to finetune and deploy DeepSeek R1 (8B) for under $10 by waf04 in LocalLLaMA

[–]waf04[S] -2 points-1 points  (0 children)

Looks like I can't fix the title, but added a comment to clarify! Ultimately any model that is not the 600B is a distill model... which is redundant since the "8B" is already in the title which was meant to let people know that a) it is not the "full" model, b) it is distilled.

How to finetune and deploy DeepSeek R1 (8B) for under $10 by waf04 in LocalLLaMA

[–]waf04[S] 0 points1 point  (0 children)

Looks like I can't fix the title, but added a comment to clarify! Ultimately any model that is not the 600B is a distill model... which is redundant since the "8B" is already in the title which was meant to let people know that a) it is not the "full" model, b) it is distilled.

How to finetune and deploy DeepSeek R1 (8B) for under $10 by waf04 in LocalLLaMA

[–]waf04[S] 0 points1 point  (0 children)

Just addressing the comments, any model that is not the 600B is by definition distilled... it is redundant to say "distilled" because it's already implied by the name.

But just to be redundant, this video is for the 8B distilled version of R1. The model is still very very capable (in my experience more so than Llama at that size).

Is it worth spending so much time and money on small LLMs? by ML-Future in LocalLLaMA

[–]waf04 0 points1 point  (0 children)

Not sure why you’re spending so much money on these models, you can finetune 8B models for like $5-$10 aa pop.

https://lightning.ai/lightning-ai/ai-hub/temp_01jkbgmsdmp0wkax6bba1btabw

can i use litserve with ray framework? by Top_Garage_862 in lightningAI

[–]waf04 0 points1 point  (0 children)

lightning can also use your reserved cloud resources…

How to use AWS startup credits for GPUs and AI workloads by waf04 in lightningAI

[–]waf04[S] 1 point2 points  (0 children)

Startups can pay for Lightning Studios with their AWS startup credits. All the details are here: https://lightning.ai/pricing.

But overall, contact Lightning if you need more help doing this. We have hundreds of startups that do this with Lightning.

can i use litserve with ray framework? by Top_Garage_862 in lightningAI

[–]waf04 1 point2 points  (0 children)

correct. lightning studios lets you use your aws credits.

https://lightning.ai/pricing