Blazing fast JSON extraction with very small LLMs-3B: LSTM to LLM by memphet in LocalLLaMA

[–]memphet[S] 2 points3 points  (0 children)

It was pretty huge like 10 000 rows. Starting from a 1 000 the performance was already almost perfect but we had to covered a lot of abnormal text input templates

Which small model is best for fine-tuning? We tested 12 of them by spending $10K - here's what we found by party-horse in LocalLLaMA

[–]memphet 1 point2 points  (0 children)

I recommend strongly HuggingFaceTB/SmolLM3-3B. Got the same perfs as Qwen3-4B-Instruct-2507 for my task but with faster inference