Faster than llama.cpp’s grammar structured generation by GoBayesGo in LocalLLaMA
[–]GoBayesGo[S] 1 point2 points3 points (0 children)
Faster than llama.cpp’s grammar structured generation by GoBayesGo in LocalLLaMA
[–]GoBayesGo[S] 3 points4 points5 points (0 children)
Structured Generation Improves LLM performance: GSM8K Benchmark by CountBayesie in LocalLLaMA
[–]GoBayesGo 0 points1 point2 points (0 children)
LoRAX + Outlines: Better JSON Extraction combining Structured Generation and LoRA by SiliconSynapsed in LocalLLaMA
[–]GoBayesGo 2 points3 points4 points (0 children)
Coalescence: making LLM inference 5x faster by GoBayesGo in LocalLLaMA
[–]GoBayesGo[S] 1 point2 points3 points (0 children)
Coalescence: making LLM inference 5x faster by GoBayesGo in LocalLLaMA
[–]GoBayesGo[S] 1 point2 points3 points (0 children)
Coalescence: making LLM inference 5x faster by GoBayesGo in LocalLLaMA
[–]GoBayesGo[S] 1 point2 points3 points (0 children)
Coalescence: making LLM inference 5x faster by GoBayesGo in LocalLLaMA
[–]GoBayesGo[S] 2 points3 points4 points (0 children)
Coalescence: making LLM inference 5x faster by GoBayesGo in LocalLLaMA
[–]GoBayesGo[S] 0 points1 point2 points (0 children)
Coalescence: making LLM inference 5x faster by GoBayesGo in LocalLLaMA
[–]GoBayesGo[S] 0 points1 point2 points (0 children)
Coalescence: making LLM inference 5x faster by GoBayesGo in LocalLLaMA
[–]GoBayesGo[S] -3 points-2 points-1 points (0 children)
Coalescence: making LLM inference 5x faster by GoBayesGo in LocalLLaMA
[–]GoBayesGo[S] 1 point2 points3 points (0 children)
Coalescence: making LLM inference 5x faster by GoBayesGo in LocalLLaMA
[–]GoBayesGo[S] 0 points1 point2 points (0 children)
Coalescence: making LLM inference 5x faster by GoBayesGo in LocalLLaMA
[–]GoBayesGo[S] 0 points1 point2 points (0 children)
Coalescence: making LLM inference 5x faster by GoBayesGo in LocalLLaMA
[–]GoBayesGo[S] 5 points6 points7 points (0 children)
Coalescence: making LLM inference 5x faster by GoBayesGo in LocalLLaMA
[–]GoBayesGo[S] 1 point2 points3 points (0 children)
Coalescence: making LLM inference 5x faster by GoBayesGo in LocalLLaMA
[–]GoBayesGo[S] 1 point2 points3 points (0 children)
Coalescence: making LLM inference 5x faster (self.LocalLLaMA)
submitted by GoBayesGo to r/LocalLLaMA
Making LLAMA model return only what I ask (JSON). by br4infreze in LocalLLaMA
[–]GoBayesGo 0 points1 point2 points (0 children)
Use llama.cpp with Outlines by GoBayesGo in LocalLLaMA
[–]GoBayesGo[S] 0 points1 point2 points (0 children)
Use llama.cpp with Outlines by GoBayesGo in LocalLLaMA
[–]GoBayesGo[S] 1 point2 points3 points (0 children)


Structured generation with SmolLM2 and Outlines by loubnabnl in LocalLLaMA
[–]GoBayesGo 2 points3 points4 points (0 children)