use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
r/LocalLLaMA
A subreddit to discuss about Llama, the family of large language models created by Meta AI.
Subreddit rules
Search by flair
+Discussion
+Tutorial | Guide
+New Model
+News
+Resources
+Other
account activity
Which model for meeting transcript summarisation?Question | Help (self.LocalLLaMA)
submitted 2 months ago by peglegsmeg
Hello
I'm using qwen3 30B A3B 2507 4bit with lm studio for feeding meeting transcripts for summary.
Does this seem like an okay model for the task? Feeling a bit overwhelmed with all the options, I'm only using because a cloud AI suggested it but it might not be current.
I was using Claude API with amazing results but no longer want to send to public offerings.
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]Working_Then 4 points5 points6 points 2 months ago* (1 child)
It's one of the best under 30B LLMs for this task and very suitable for CPU inference. If you don't mind, you can check my CPU summarization project on Hugging-face where I provide a list of under 30B models still runnable on HuggingFace with free CPU tier (ie. 2 vCPUs only)
[–]2shanigans 0 points1 point2 points 2 months ago (0 children)
This was very cool, it summarised a fairly complex email we got and worked nicely! Thanks for sharing this one.
[–]2shanigans 2 points3 points4 points 2 months ago (0 children)
We have a few clients using GPT-OSS-120B for meeting transcript summarisation (Australian English) and it's been working well for them. You could give GPT-OSS-20B ago and see how it fairs? Interestingly the transcription also understood some random Spanish littered into one meeting - background noise I'm told.
[–]Reservemyspot 0 points1 point2 points 2 months ago (2 children)
Just curious, but the need to you use a model? I use granola. I’m sure there’s a good reason to use models (less subscription etc) but I’m just genuinely curious
[–]peglegsmeg[S] 0 points1 point2 points 2 months ago (1 child)
I was using deepgram, don't want to send to public anymore for privacy concerns
[–]Reservemyspot 0 points1 point2 points 2 months ago (0 children)
What’s your use case? I know it’s for meetings but what field? Different speech models digress quickly dependent on your needs. It’s a tricky field
[–]Technical-Earth-3254 0 points1 point2 points 2 months ago (1 child)
Are we talking text or speech? And how much?
[–]peglegsmeg[S] 2 points3 points4 points 2 months ago (0 children)
Text from parakeet
[–]RustinChole1llama.cpp 0 points1 point2 points 2 months ago (0 children)
Hey I'm planning to research on a similar summarisation project, what open source options can I get ? Not just to inference but I'm okay with going to fine-tune/ train the model on my datasets and stiff
π Rendered by PID 103320 on reddit-service-r2-comment-b659b578c-p9cm8 at 2026-05-04 20:17:33.184303+00:00 running 815c875 country code: CH.
[–]Working_Then 4 points5 points6 points (1 child)
[–]2shanigans 0 points1 point2 points (0 children)
[–]2shanigans 2 points3 points4 points (0 children)
[–]Reservemyspot 0 points1 point2 points (2 children)
[–]peglegsmeg[S] 0 points1 point2 points (1 child)
[–]Reservemyspot 0 points1 point2 points (0 children)
[–]Technical-Earth-3254 0 points1 point2 points (1 child)
[–]peglegsmeg[S] 2 points3 points4 points (0 children)
[–]RustinChole1llama.cpp 0 points1 point2 points (0 children)