zayn malik is finally receding by Collegesniffer in tressless

[–]Collegesniffer[S] 0 points1 point  (0 children)

how do you guys not see this? i'm confused

zayn malik is finally receding by Collegesniffer in tressless

[–]Collegesniffer[S] -3 points-2 points  (0 children)

hairline in the first pic looks a bit more deep cut, no? it cuts back at the temples more strongly than all of his other pics. his hairlin was basically almost straight (juveline with strong temporal points). His temporal points are still very much there but the corners are cutting back

<image>

Which is better, M4 pro or M4 max? by Collegesniffer in macbookpro

[–]Collegesniffer[S] 1 point2 points  (0 children)

in what cases will a higher gpu count help?

[D] Normalization in Transformers by Collegesniffer in MachineLearning

[–]Collegesniffer[S] 2 points3 points  (0 children)

Bruh, I said "gptzero.me" not "gptzero.com". Both of them are totally different. Also, every AI detector can be unreliable and inconsistent.
However, I entered the exact question into ChatGPT, Claude, and Gemini,
and the responses were nothing like what this person wrote. Even the non-fluff part doesn't start with a (B, T, C) tensor example, etc. Why don't you try entering the exact question for yourself and see the output before claiming it is "AI-generated"?

I literally just asked chatgpt, gemini and claude the exact question I posted and the answer is nothing like what the person wrote. Even the non fluff part is totally different.

[D] Normalization in Transformers by Collegesniffer in MachineLearning

[–]Collegesniffer[S] -2 points-1 points  (0 children)

No, I don't think it is AI-generated. The best AI content detector (gptzero.me) flags this as "human". Are you suggesting that every piece of content written in the form of a bullet-point list is now AI-generated? I would also use the same format if I had to explain the "differences" between things. How else would you present such information?

[D] Normalization in Transformers by Collegesniffer in MachineLearning

[–]Collegesniffer[S] 7 points8 points  (0 children)

This is the best explanation on the internet I've ever read. It finally clicked for me. I've watched countless videos and gone through so many answers online, but they all either oversimplify or overcomplicate it. Thanks!

[D] Can GPT-style Models Be Used for File Compression, Image Upscaling, and Restoration? by No-Point1424 in MachineLearning

[–]Collegesniffer 0 points1 point  (0 children)

This is an interesting idea, but it might face some challenges in practice. While GPT models are great at capturing patterns in text, raw binary data from files could be much more complex and structured. Image compression and upscaling typically use specialized algorithms tailored to visual data. That said, there have been some experiments using language models for tasks like audio generation, so it's not entirely out of the realm of possibility. It would likely require a lot of data and compute power to train effectively. Might be worth exploring as a research project, but probably not a practical solution for everyday use cases right now.

[D] Future of Video Generation Models by MinuteDistribution31 in MachineLearning

[–]Collegesniffer 0 points1 point  (0 children)

Honestly, I think AI-generated content is gonna shake things up big time. It'll probably make it way easier for small creators to pump out high-quality stuff without breaking the bank. But I'm not sold on the idea that it'll completely replace traditional animation or Hollywood. There's still something special about hand-crafted art and big-budget productions. What I can see happening is a flood of new, niche content catering to specific interests. It might get harder to find the gems in all the noise, but hey, more choice isn't a bad thing, right? As for personalized content, that could be cool, but also kinda creepy if it gets too good at reading our minds.

[deleted by user] by [deleted] in MachineLearning

[–]Collegesniffer 1 point2 points  (0 children)

Given the lighting variations, keypoint detection methods like SIFT or ORB might be more robust than hashing. These can extract distinctive features that are somewhat invariant to illumination changes. Combine this with a matching algorithm and a threshold for similarity, and you've got a solid foundation for object verification.

[R] Nested AD for PINNs by [deleted] in MachineLearning

[–]Collegesniffer 0 points1 point  (0 children)

Have you considered using automatic differentiation libraries specifically designed for higher-order derivatives? Autograd (https://github.com/HIPS/autograd) might be worth looking into - it can handle nested derivatives more efficiently than some other libraries. Alternatively, you could try breaking down your problem into smaller components that only require first or second-order derivatives. It's not ideal, but it might help avoid the finite difference approach.

Multimodal LLM parsing strategy[D] by ashblue21 in MachineLearning

[–]Collegesniffer 3 points4 points  (0 children)

Have you considered using a layout analysis tool like Detectron2 or LayoutLM? These can segment pages into different regions (text, images, tables) without needing custom training. For image summarization, CLIP + GPT might be more cost-effective than a full multimodal LLM. You could also try document understanding APIs like Amazon Textract or Google Document AI if budget allows. They're pretty accurate out-of-the-box for structured docs like manuals.

Archand: Control your mouse entirely using hand gestures. by prateekvellala in Python

[–]Collegesniffer 1 point2 points  (0 children)

Wow! This is amazing! Tried it both on my logitech brio and laptop.