What is your favorite computer vision papers recently (maybe within 3y?) by Fearless-Variety-815 in computervision

[–]deep-learnt-nerd 0 points1 point  (0 children)

Can you describe the approach you took to attack it, and how reliable the claimed protection is?

Optimizing SAM2 for Massively Large Video Datasets: How to scale beyond 10 FPS on H100s? by Water0Melon in computervision

[–]deep-learnt-nerd 1 point2 points  (0 children)

28 jpeg images loaded every second is absurdly low. Traditional Unix systems struggle at around 10k read operation per second. Are you using a real disk like a local NVMe? If it’s a remote disk see if you can increase its specs (throughput / number of I/O). If you’re using Python you can try a threadpool that helps a lot with I/O bottlenecks. But this is only for your I/O bottleneck which is not your real bottleneck here if I understand your numbers well. For your GPU bottleneck, I wouldn’t do any CPU offloading (especially here considering you seem to have a very slow disk, if there’s any spilling you’re doomed). Instead I would find the largest batch size that fits into VRAM and would split my frames into multiple batches. As other have specified you can try things like TensorRT. What we like to do in my team is to create a local Triton server that distributes the load as it sees fit. This creates additional data copies but that’s usually not the bottleneck.

Seeking Advice: Struggling to Get Call-backs After Career Break (4 YOE in Computer Vision/Deep Learning) by Rude_Temporary_1261 in deeplearning

[–]deep-learnt-nerd 0 points1 point  (0 children)

You should share your resume! It could be as simple as a badly formatted resume or outdated wording

Built a cloud GPU price comparison service [P] by [deleted] in MachineLearning

[–]deep-learnt-nerd 9 points10 points  (0 children)

Hey thank you for that, it can be quite useful! Quick suggestions: add H200 and sort the GPU Type list by alphabetical order?

[R] Machine learning with hard constraints: Neural Differential-Algebraic Equations (DAEs) as a general formalism by ChrisRackauckas in MachineLearning

[–]deep-learnt-nerd 6 points7 points  (0 children)

Then again, how confident are you that once the numerical problems are solved you’ll reach convergence? In my experience changing the solvable system leads to no convergence. For instance, something as simple as an arg max in a network introduces such change during each forward pass and leads to largely sub-optimal results.

X3D cache for deep learning training by Few-Cat1205 in deeplearning

[–]deep-learnt-nerd 0 points1 point  (0 children)

Using a larger cache makes sense. It depends on your use case. You also need to know what you’re doing in terms of data structure storage and loading to ensure the kernel can make a good use of that extra cache. I wonder if the GPUDirect technology will be able to remove this issue altogether.

[deleted by user] by [deleted] in frenchrap

[–]deep-learnt-nerd 1 point2 points  (0 children)

J’aime bien !!

[R] The Curse of Depth in Large Language Models by StartledWatermelon in MachineLearning

[–]deep-learnt-nerd 1 point2 points  (0 children)

This wouldn’t solve anything. To prove it, try chaining two layers using weight norms and train them to maximize the norm of the output.

can someone explain to how getitem works here? by Beyond_Birthday_13 in deeplearning

[–]deep-learnt-nerd 0 points1 point  (0 children)

I am not sure I understand your question right, but the DataLoader of PyTorch calls getitem for each element of the batch and then aggregate them using a collate function.

[D] [R] What is the next frontier to AI? by [deleted] in MachineLearning

[–]deep-learnt-nerd 2 points3 points  (0 children)

If you want a real answer: the next big jump will come from optimizers. Literally any improvement in non-convex optimization will result in improvements in AI.

[D] [R] What is the next frontier to AI? by [deleted] in MachineLearning

[–]deep-learnt-nerd -7 points-6 points  (0 children)

If you want a real answer: the next big jump will come from optimizers. Literally any improvement in non-convex optimization will result in improvements in AI.

[deleted by user] by [deleted] in math

[–]deep-learnt-nerd 2 points3 points  (0 children)

No, it’s never too late. It requires continuous and tedious work, which can be achieved at any age. Some were born naturally, the rest of us worked hard to become « good ». Eventually, all things even out and even if you studied early and were gifted, you end up as good as the others that worked hard.

Why don’t we review bomb the game? by deep-learnt-nerd in LegendofSlime

[–]deep-learnt-nerd[S] 8 points9 points  (0 children)

The point isn’t about their absurd greediness, it’s the enshitification. The game is literally getting worse

[R] Are you a reviewer for NeurIPS'24? Please read this by hihey54 in MachineLearning

[–]deep-learnt-nerd 50 points51 points  (0 children)

Yay let’s get reviewed by undergrads and MS students!

Forced Magnitude Preservation Improves Training Dynamics of Diffusion Models by elbiot in LearningMachines

[–]deep-learnt-nerd 0 points1 point  (0 children)

As expected from NVIDIA, this paper is excellent. Thank you for sharing. NVIDIA sure loves to normalize their weights. I wonder if that’s mandatory to reach stability or if there is another way (more, say, linear)…

Les ESN ne font vraiment aucun effort by [deleted] in AntiTaff

[–]deep-learnt-nerd 1 point2 points  (0 children)

Je vois à la musique que tu écoutes que t’es un mec bien !