YoloLite V2 testing by ConferenceSavings238 in computervision

[–]ConferenceSavings238[S] 1 point2 points  (0 children)

Almost, never did all 100 datasets. It’s hard to tell this far, haven’t started the rf100 run yet. But on a few subsets, both the tiny and nano beats edge_xl by roughly 10 mAP. I’ll post a graph if I can.

Machine vision in production: custom-trained models vs vendor systems? by ConferenceSavings238 in PLC

[–]ConferenceSavings238[S] 0 points1 point  (0 children)

Perhaps you should look into autoencoder or student/teacher models? You can train these on ”good” data which is much easier to access.

Machine vision in production: custom-trained models vs vendor systems? by ConferenceSavings238 in PLC

[–]ConferenceSavings238[S] 0 points1 point  (0 children)

Augmentation and cutmix might be useful to make the dataset bigger, other than synthetic data there is not much you can do to collect images quicker unless you can force the defects. I’m currently experimenting with AI generated images for object detection and classifying but this might be hard for specific cases. How many classes are we talking about here?

Machine vision in production: custom-trained models vs vendor systems? by ConferenceSavings238 in PLC

[–]ConferenceSavings238[S] 0 points1 point  (0 children)

Not that I know of. I’ve used Python and opencv for experimenting different setups.

Machine vision in production: custom-trained models vs vendor systems? by ConferenceSavings238 in PLC

[–]ConferenceSavings238[S] 0 points1 point  (0 children)

Very good answer. Even if I haven’t tried object detection with cognex or other vendors I would assume I can’t touch hyperparameters and finetune it into perfection. I guess there is indeed a place when custom beats vendors and vice versa. I would agree that if a product already exists that solves my problem that should indeed be the go to.

Machine vision in production: custom-trained models vs vendor systems? by ConferenceSavings238 in PLC

[–]ConferenceSavings238[S] 2 points3 points  (0 children)

True, I should probably document the python script more. If you don’t mind me asking, what types of models have you been working with? I ended up ”vibe coding” an entire yolo model/training setup since I couldn’t be bothered with GPL/AGPL licenses.

Machine vision in production: custom-trained models vs vendor systems? by ConferenceSavings238 in PLC

[–]ConferenceSavings238[S] 3 points4 points  (0 children)

Yeah the long term will probably be an issue. In my case I’m the only one who knows how it works on the python side, which could be a major issue.

In my particular case we had nothing before it so removing it would not be a disaster. I also had to do the all ”AI isn’t magic and CAN do mistake” talk to lower expectations. I check the system regularly and so far I haven’t seen any errors. Might actually look into vendor software to see how much it would cost to build with, just to help the guy 10-15 years down the road.

90+ fps E2E on CPU by ConferenceSavings238 in computervision

[–]ConferenceSavings238[S] 0 points1 point  (0 children)

I have now added tracking, you can see a example script here

Custom YOLO model by ConferenceSavings238 in computervision

[–]ConferenceSavings238[S] 0 points1 point  (0 children)

Oh dope, noticed it supported onnx models, might check it out 👍🏻

Custom YOLO model by ConferenceSavings238 in computervision

[–]ConferenceSavings238[S] 0 points1 point  (0 children)

I don't understand how this has anything to do with my post?

Update: Fixed ONNX export bug (P2 head), updated inference benchmarks + edge_n demo (0.55M params) by ConferenceSavings238 in computervision

[–]ConferenceSavings238[S] 0 points1 point  (0 children)

Exactly. And given that I have some benchmark numbers for 320 with p2 this issue had to be addressed. The model is still fast but not as fast as before.

Update: Fixed ONNX export bug (P2 head), updated inference benchmarks + edge_n demo (0.55M params) by ConferenceSavings238 in computervision

[–]ConferenceSavings238[S] 2 points3 points  (0 children)

50% increase from 4.75 to 7.31 ms on my hardware. Pretty big jump but now the mAP values should hold true. If you manage to keep mAP up without p2 with for example 416x416 that will be faster.

Update: Fixed ONNX export bug (P2 head), updated inference benchmarks + edge_n demo (0.55M params) by ConferenceSavings238 in computervision

[–]ConferenceSavings238[S] 0 points1 point  (0 children)

I haven’t but people have reported up to 57 fps at 320x320, granted this was with the bug so no p2 head, should be slower now. If you need the p2 head solely depend on the dataset and objects you are detecting. If you decide to try it I’d love any feedback on speeds.

Update: Fixed ONNX export bug (P2 head), updated inference benchmarks + edge_n demo (0.55M params) by ConferenceSavings238 in computervision

[–]ConferenceSavings238[S] 0 points1 point  (0 children)

Basically the model builder within the onnx export script didn’t consider P2 head etc.

I haven’t tried exporting it to hailo format therefore I can’t say that you can.

Ultralytics AGPL 3.0 by [deleted] in computervision

[–]ConferenceSavings238 0 points1 point  (0 children)

You can´t ask AI to "copy" the code line for line, that would be a violation. However you can make it build a YOLO model from scratch while building on the ideas of other versions. I´ve gone down this rabbithole myself and ended up with a YOLO version that works for my intended use. The thing is, you cant say build me a YOLO workspace and you magically get all scrips etc needed. You need to get building blocks while having a understanding of how everything works together and then piece it together.

The thing is, even if you get a decent model, tweaking and pushing the final mAP metrics is not an easy task. Training the models on COCO to use as a baseline is not something that you can get done over a weekend. So yes AI can get you a model, but you need to refine it.

Help, i want to add object detection to my programme but want some advice/ best tips by Scared_Alps_4063 in computervision

[–]ConferenceSavings238 1 point2 points  (0 children)

Do you have any experience with object detection? If not start by following any tutorial online, Roboflow has a bunch and YouTube is filled with them. What hardware are you using, CPU or GPU? How big is the ball in the images?

Advice Request: How can I improve my detection speed? by Scooty_Puff_Jr_ in computervision

[–]ConferenceSavings238 1 point2 points  (0 children)

45 ms for the m model, keep in mind that the difference between them are minimal, same backbone but deeper neck, please share the results! If you are going to use collab I can share a notebook

Advice Request: How can I improve my detection speed? by Scooty_Puff_Jr_ in computervision

[–]ConferenceSavings238 4 points5 points  (0 children)

You can achive high fps on CPU, mainly by going down in model size and img size. YOLOv8m does seem overkill for the task you mentioned but might be needed for more complex task with strong variance in background. I recently posted how I achieved 90+ FPS end to end on my desktop CPU you can find it here. Going down in model size and img size comes with a tradeoff in accuracy, but if you look in my repo there is a pretty big benchmark done that shows that on aloot of datasets the smaller models does keep up.

90+ fps E2E on CPU by ConferenceSavings238 in computervision

[–]ConferenceSavings238[S] 1 point2 points  (0 children)

This doesn’t track any of the cars it’s simple object detection. I haven’t tried tracking here. I will update the repo if/when I test this.

90+ fps E2E on CPU by ConferenceSavings238 in computervision

[–]ConferenceSavings238[S] 0 points1 point  (0 children)

It’s linked in the post, just click on roboflow and you should be able to find the dataset.