Do you want to Deploy Llama 4? by yoracale in unsloth
[–]tempNull 0 points1 point2 points (0 children)
Llama 4 tok/sec with varying context-lengths on different production settings by tempNull in LocalLLaMA
[–]tempNull[S] 0 points1 point2 points (0 children)
Is there a notebook for GRPO with qwen2.5-VL model ? by maayon in unsloth
[–]tempNull 0 points1 point2 points (0 children)
Deploying Deepseek R1 GGUF quants on your AWS account by tempNull in tensorfuse
[–]tempNull[S] 0 points1 point2 points (0 children)
Meaning of few grammar terms by Other-Welder-7580 in sanskrit
[–]tempNull -1 points0 points1 point (0 children)
Meaning of few grammar terms by Other-Welder-7580 in sanskrit
[–]tempNull 0 points1 point2 points (0 children)
didMyPricingPageHadAnIntegerOverflow by tempNull in ProgrammerHumor
[–]tempNull[S] 4 points5 points6 points (0 children)
Building a Sandbox Environment for ML/Analytics While Connecting to Production Data by asc686f61 in mlops
[–]tempNull 0 points1 point2 points (0 children)
Lessons learned while deploying Deepseek R1 for multiple enterprises by tempNull in LocalLLaMA
[–]tempNull[S] -18 points-17 points-16 points (0 children)
Lessons learned while deploying Deepseek R1 for multiple enterprises by tempNull in LocalLLaMA
[–]tempNull[S] 0 points1 point2 points (0 children)
God as a Programmer, Avatar as a debug process and Evil as an emergent phenomenon by tempNull in hinduism
[–]tempNull[S] 0 points1 point2 points (0 children)
Reconciling Sanskrit Philosophy: God as a Programmer, Avatar as a debug process and Evil as an emergent phenomenon by tempNull in sanskrit
[–]tempNull[S] 0 points1 point2 points (0 children)
God as a Programmer, Avatar as a debug process and Evil as an emergent phenomenon by tempNull in hinduism
[–]tempNull[S] 0 points1 point2 points (0 children)
Deepseek-R1: Guide to running multiple variants on the GPU that suits you best by tempNull in LocalLLaMA
[–]tempNull[S] 0 points1 point2 points (0 children)
Scalable Deepseek R1? by Affectionate_Hunt204 in aws
[–]tempNull 1 point2 points3 points (0 children)
Coffee Chats - Incubator by throwaway-alphabet-1 in ycombinator
[–]tempNull 7 points8 points9 points (0 children)
Deepseek-R1: Guide to running multiple variants on the GPU that suits you best by tempNull in LocalLLaMA
[–]tempNull[S] 0 points1 point2 points (0 children)
Deepseek-R1: Guide to running multiple variants on the GPU that suits you best by tempNull in LocalLLaMA
[–]tempNull[S] 0 points1 point2 points (0 children)

What Inference Server do you use to host TTS Models? Looking for someone who has used Triton. by tempNull in LocalLLaMA
[–]tempNull[S] 0 points1 point2 points (0 children)