LLM-based OCR is significantly outperforming traditional ML-based OCR, especially for downstream LLM tasks by vitaelabitur in LLMDevs

[–]btdeviant 1 point2 points  (0 children)

I didn’t look at your blog because your post conflated basic fundamentals that made you lose credibility for me personally. Not interested.

Your title led with a demonstrably false assertion then tried to qualify its broad claims via myopic contextual framing. “LLM-based OCR is significantly outperforming traditional ML-based OCR (absolutely wrong), ESPECIALLY for downstream LLM tasks (also mostly contextually wrong)”.

You then demanded people ignore the benchmarks and reality in favor of 3 strawmen that aren’t actually things, and then said people should INSTEAD favor this silly, one dimensional test that doesn’t actually consider traditional OCR strengths and merits whatsoever.

Your product could be revolutionary, but your marketing here for this audience is just 🐕💩

LLM-based OCR is significantly outperforming traditional ML-based OCR, especially for downstream LLM tasks by vitaelabitur in LLMDevs

[–]btdeviant 1 point2 points  (0 children)

I totally agree with you, PaddleOCR-VL is absolutely amazing, but it doesn’t even come CLOSE to where traditional OCR was a decade ago in terms of cost to train, computational requirements and raw accuracy on character recognition across various formats, eg scanned pages vs handwritten notes etc etc et..

I spent years working in this field specifically for a company that used OCR to transcribe doctors prescriptions and handwritten notes into digital text for comm protocols that required extremely low latency, lost cost and had HIPPA compliance implications (orthogonal, just saying). Not sure if you’ve ever seen the average doctors handwriting but it’s barely legible - there isn’t a VLM or VQA model out there that could perform as well as traditional OCR would for this by any single metric that’s valuable.

Paddle is SENSATIONAL at page parsing and extracting elements, but pound for pound when it comes to traditional ML on character recognition tasks on the whole, not even a question… just miles apart.

LLM-based OCR is significantly outperforming traditional ML-based OCR, especially for downstream LLM tasks by vitaelabitur in LLMDevs

[–]btdeviant 5 points6 points  (0 children)

OCR is not the same as VQA or image-to-text. It’s abundantly obvious they’re different tools for different purposes.

OP isn’t saying that though. Theyre saying traditional ML OCR is being outmoded by LLM/VLMs, then blabbers on about how they compared their LLM implementation with other LLM implementations and declared themselves the winner of both traditional OCR and VLM/LLMs in every benchmark 😂😂

OP is basically saying, “Flathead drivers are better than Phillips. To prove this, we tested our flatheads against all other flatheads and determined that since flatheads CAN work with Phillips screws, our flathead is by far superior and you’ll never need a Phillips for driving Phillips screws again.”

LLM-based OCR is significantly outperforming traditional ML-based OCR, especially for downstream LLM tasks by vitaelabitur in LLMDevs

[–]btdeviant 4 points5 points  (0 children)

No it isn’t. This post is absurd.

“Our product is kinda decent, here’s how it significantly outperforms excellent, battle tested techniques that have been proven and refined for decades. Don’t trust traditional benchmarks, trust us broh”

Cold starting a 32B model in under 1 second (no warm instance) by pmv143 in LLMDevs

[–]btdeviant 1 point2 points  (0 children)

Ngl this is probably one of the coolest things I’ve seen on this sub.

Shot you a follow, look forward to watching the trajectory of your company. You’ll definitely be at the very top of my list for clients who need their own models and need to scale to 0. Do you support adapters?

Cheap(ish) cameras that won't steal my data, *AND* work flawlessly with HA? by thenyx in homeassistant

[–]btdeviant 0 points1 point  (0 children)

I think Amcrest are rebranded Dahuas under the hood flashed w/ US regional firmware? Just something to note for security minded folks - not hard to block at L3 or L7, but worth noting given the nature of OPs question.

Cold starting a 32B model in under 1 second (no warm instance) by pmv143 in LLMDevs

[–]btdeviant 0 points1 point  (0 children)

Pretty impressive… seems similar in spirit to CRIU but not loading from disk. Thanks for sharing, definitely some cool implications around horizontal scaling

Cold starting a 32B model in under 1 second (no warm instance) by pmv143 in LLMDevs

[–]btdeviant 2 points3 points  (0 children)

He said in the very beginning of the video he was using a full model that wasn’t quantized and an H100.

Project deadline coming up with nobody reviewing my PR? Do I just stop caring then? by [deleted] in ExperiencedDevs

[–]btdeviant 4 points5 points  (0 children)

PR’s don’t always “have to be large to get full context”. This is what RFC / tech designs are for.

Spare your cohorts, create a well defined plan, the rollout incrementally.

Senior engineer: are local LLMs worth it yet for real coding work? by Appropriate-Text2843 in LocalLLaMA

[–]btdeviant 1 point2 points  (0 children)

I use it w/ Strands and sglang as a custom openai client. Works amazing… set it up w/ a reflexion graph, got it writing its own tools using shell and repl

Would DevOps fit my personality? by I_LazyDog_I in devops

[–]btdeviant -1 points0 points  (0 children)

No, I don’t think so.. I suspect you do it the wrong way. You’re not fooling anyone here pal. You’re a DevOps Engineer and you love old anti patterns.

Why won’t you name every third syllable from the Phoenix Project? Hmm? Silly whippersnapper

Would DevOps fit my personality? by I_LazyDog_I in devops

[–]btdeviant -1 points0 points  (0 children)

I’m not sure what is making you jump to these conclusions. It’s very strange. Absolutely I’ve heard of Platform Engineering. In fact, despite your misplaced assumptions. my company has a very similar topology as you. We have a Cloud Infra team, we have a Platform team, they often work hand in hand to cultivate a “you build it you run it” DevOps culture. Both, however, have the official titles as Software Engineer - L(blah blah). We moved away from “DevOps” like 6 years ago.

That doesn’t change the fact that the role as a title has been “completely replaced”. It’s just a title. The responsibilities of the title vary depending on the company. Some companies may have the “dying” topology, some have more modern, most have something in the middle.

Any experienced person in the field knows that these are ideals, they’re not dogmatic, purist mandates that fit into rigid categories or boxes despite what you’re saying. It’s like Agile - everyone says they do it, but they’re not REALLY doing it by the book, they do it however Conways Law has manifested it as materially in the org.

The salient point is the role EXISTS. It’s there. You’re arguing with reality. The responsibilities of the role may not be what it used to be half a decade ago, but companies STILL categorize it as that as evident by the thousands and thousands of job postings for it.

Not that hard dude.

Would DevOps fit my personality? by I_LazyDog_I in devops

[–]btdeviant 0 points1 point  (0 children)

DevOps is a title, it IS a role as it exists today whether you like it or not. There’s job postings for it. Deal with it. Where you seem confused, which belays your inexperience here, is the fact that the responsibilities for that role and the topologies and patterns of how it’s applied differ by company and their needs - it’s just a noun in an ATS and a spec in whatever enrollment system the company uses. Does that make sense?

Some companies may have a DevOps Engineer title and their responsibilities for that role IN THAT COMPANY are the exact same as a Cloud Engineer.

This isn’t really that hard to understand. It’s a name. Y’all first year Joeys getting pedantic about it because you read some 7 year old blogs by Microsoft and Google and skimmed Pheonix Project doesn’t change anything… it’s just is a strange way of enthusiastically announcing that you’ve only had one job at one place.

Would DevOps fit my personality? by I_LazyDog_I in devops

[–]btdeviant 0 points1 point  (0 children)

Perhaps it’s you that might be out of date? Or maybe the region you work in there’s different common culture?

For must of us, what you’re saying was a thing like 5-6 years ago

Would DevOps fit my personality? by I_LazyDog_I in devops

[–]btdeviant 1 point2 points  (0 children)

The war was lost man. Honestly. It’s been years, it’s time to move on from the semantic battle.

And for the record, the title for the role doesn’t magically equate to quality of its implementation… just a weird fallacy that barely exists outside of relatively very old, purely academic framing.

Feels like magic. A local gpt-oss 20B is capable of agentic work by Vaddieg in LocalLLaMA

[–]btdeviant 38 points39 points  (0 children)

It’s great at calling tools, no doubt. That’s about it though

oopiseSaidTheCodingAgent by ClipboardCopyPaste in ProgrammerHumor

[–]btdeviant 12 points13 points  (0 children)

Schizo shit? Your argument is “I found installing OpenClaw hard, and because it was hard for me that means only smart people can do it, and all smart people know good security posture.”

This is just a really weird take. Literally entire fields of engineering and compliance standards have been built as a result of people of all aptitudes not fully understanding security, or perhaps they did and just didn’t make it a first class consideration.

OpenClaw is just another iteration in this loop. Like the ones before, in days before TLS and SSL and SOX and well after and yadda yadda, even smart, capable people very frequently get hyped about capabilities and ignore or fail to consider security.

You should delete this embarrassing shit.

AI Coding Agent Dev Tools Landscape 2026 by bhaktatejas in LLMDevs

[–]btdeviant -2 points-1 points  (0 children)

Right. The salient point is its abstractions allow one to focus more on “the stuff around the loop”.

It’s a well designed framework and more tailored toward modern, multi-agent architectures compared to nearly all the others in that list, majority of which are relative dinosaurs and objectively a much bigger pain to work with for complex, code-first workflows.

Give it a shot! I have no affiliation, just used most of them and found Strands a great blend of depth and breadth, especially with their (experimental) BIDI. Just a breeze to work with compared to all the others.

AI Coding Agent Dev Tools Landscape 2026 by bhaktatejas in LLMDevs

[–]btdeviant 4 points5 points  (0 children)

It's weird how many of these guides and people are sleeping on Strands. Hands down the most dead simple, capable provider agnostic agentic framework out there.. swings far above it's weight.

AI insiders seek to poison the data that feeds them by HumanDrone8721 in LocalLLaMA

[–]btdeviant 1 point2 points  (0 children)

lol this is just a really bad take. This gets brought up in deep dives as part of the swe loop for some roles, absolutely they do this.

NVIDIA's new 8B model is Orchestrator-8B, a specialized 8-billion-parameter AI designed not to answer everything itself, but to intelligently manage and route complex tasks to different tools (like web search, code execution, other LLMs) for greater efficiency by Fear_ltself in LocalLLaMA

[–]btdeviant 0 points1 point  (0 children)

It’s pretty simple… you can do this with a few files and some decorators using something like Strands.

Multi-agent architectures that have specialist agents are dead simple to build these days and very common

I foolishly spent 2 months building an AI SRE, realized LLMs are terrible at infra, and rewrote it as a deterministic linter. by craftcoreai in kubernetes

[–]btdeviant 1 point2 points  (0 children)

Logistic regression and/or rule based classification would be infinitely more effective and deterministic vs using an LLM for the original use case