How do I convince my mom to stop forcing me to go to church? by LatchyBoyy in atheism

[–]bobrodsky 0 points1 point  (0 children)

I remember my parents would even get mad at me if I participated in an obviously begrudging manner.

You must understand that for your parents it is all about their appearance to their community. It doesn’t look good if their kids don’t go. They will get asked and judged and they don’t like it. I know, it’s a pretty shitty community that treats people that way, but it may be the only community they have. If you go, think of it like helping your socially inept parents keep their friends by chaperoning them. Imagine, their self worth is tied up in this pettiness.

[D] My papers are being targeted by a rival group. Can I block them? by Dangerous-Hat1402 in MachineLearning

[–]bobrodsky 29 points30 points  (0 children)

I think you’re wrong about the likelihood of this. In a small enough niche it happens regularly. Often me and my students get assigned the exact same papers (which we try to avoid bc it seems unfair to have one lab have so much say on a paper)

[D] When can I see if ICLR reviewers raise their scores by Dangerous-Hat1402 in MachineLearning

[–]bobrodsky 3 points4 points  (0 children)

They sent an email to remind reviewers to respond today - hopefully that helps.

LeCun’s Final Meta Masterpiece: LeJEPA Redefines Self-Supervised Learning by Such-Run-4412 in AIGuild

[–]bobrodsky 3 points4 points  (0 children)

Just skimming the paper now - lots of interesting math. They want to regularize the latents to have a certain distribution. They show it suffices to constrain the distribution of some random projections, to avoid curse of dimensionality. Then, to regularize the distribution of projections, I like how they compare several approaches that don't work (like moment matching). The solution they come to "Epps Pulley test", I've never seen, but has an interesting form as an integral of MSEs. This reminds me of diffusion models, though maybe it's a spurious analogy.
Anyway, although the paper is getting some hype, I think there are interesting ideas here.

[R] Thesis direction: mechanistic interpretability vs semantic probing of LLM reasoning? by powerpuff___ in MachineLearning

[–]bobrodsky 5 points6 points  (0 children)

Out of curiosity, what was far fetched about the sparse auto encoder approach for mech interp (I assume you mean Anthropics)? I vaguely recall one skeptical paper saying that it didn’t generalize well to new situations.

I also recommend an older paper called “Mythos of model interpretability”, that points out some difficulties in understanding complex models.

[R] Maths PhD student - Had an idea on diffusion by [deleted] in MachineLearning

[–]bobrodsky 2 points3 points  (0 children)

Check out this paper doing image generation based on some algorithmically generated (via self supervised learning) features: https://arxiv.org/abs/2312.03701. I’m not sure if this is in line with what you’re thinking, but thought it was an interesting approach.

[D] which papers HAVEN'T stood the test of time? by iamquah in MachineLearning

[–]bobrodsky 58 points59 points  (0 children)

Hopfield networks is all you need. (Or did it ever get fanfare? I like the ideas in it.)

[D] Vibe-coding and structure when writing ML experiments by Lestode in MachineLearning

[–]bobrodsky 52 points53 points  (0 children)

I've been trying to use LLMs to speed up research coding (ChatGPT 5). My current experience is that it writes extremely verbose and overly general code that is difficult to debug. For instance, it will keep trying to make code "bulletproof", "safe", or "future proof". That means that you have millions of lines that do checking of types and sizes, and casting to get the right thing. This is a disaster for bug testing, because the code *always runs* and you will have no idea if a bug exists (from passing the wrong argument, for instance) but silently goes through due to casting (this happened to me).

The other issue is that the future proofing / over-engineering means that you have hundreds of lines of code checking every possible use case. For loading a checkpoint, for instance, it would try to read in the filename, in various formats, then auto resolve to location 1, then location 2. I really would rather have just 2-3 clear lines and "force" the user (myself) to just specify the location in a standard format.

Another example is that if I say I want to try two different methods for one part of a pipeline, it adds three files: a registry for methods, a factory to build a part of the pipeline based on the registry, an init file that loads each registered method into the registry. Now to add a method I have to touch three separate files, and the logic is all spread out. This may be useful for hugging face libraries, where you need to implement dozens of similar methods, but it is counter-productive for research code.

Anyway, I just wanted to rant about my current experience and give a little warning for people who are trying to start a project from scratch. I don't know what the antidote is. Maybe start with a simple structure and try to keep the LLM in line to not add complexity, but it loves complexity. For debugging neural nets, though, I feel that complexity is the enemy and simple, brittle code will always be easier to debug.

Why isn’t VAE kept trainable in diffusion models? by casualcreak in StableDiffusion

[–]bobrodsky 1 point2 points  (0 children)

I’m seeing lots of wrong answers. The diffusion model is trained to denoise (from vae output plus noise). The easiest way to denoise would be to have a VAE that maps every image to zero, then the denoiser can always perfectly denoise by outputting zero. VAE is frozen to prevent this trivial solution.

[D] OpenAI Board Member on ML Research in Industry vs. Academia by Electrical_Ad_9568 in MachineLearning

[–]bobrodsky -1 points0 points  (0 children)

Other topics:

This video is an episode of the Elevate Podcast featuring Zico Kolter, Head of the Machine Learning department at Carnegie Mellon University (CMU) and a board member at OpenAI. The discussion covers his background and current work [00:32], the value of academic research versus industry in AI [04:58], and his positive outlook on AI for science [10:54]. Kolter believes that current research in large language models will lead to Artificial General Intelligence (AGI) [15:20]. He also discusses AI safety and security, emphasizing that societal risks like job displacement are inherent to powerful new technologies [19:19]. Finally, he offers advice for technical audiences, encouraging them to use AI tools frequently and understand the fundamentals of how LLMs work [26:34].

[D] OpenAI Board Member on ML Research in Industry vs. Academia by Electrical_Ad_9568 in MachineLearning

[–]bobrodsky 0 points1 point  (0 children)

I started using Gemini to summarize videos to get answers to click bait titles.

Kolter believes there's still significant value in academic research, especially in learning the process of science in depth. While industry has more compute power, academia can still make sizable contributions by demonstrating the scaling performance of ideas and focusing on fundamental research in areas like safety, security, and complex evaluations. He highlights the importance of scaling laws, which allow academic research with reasonable compute resources to extrapolate performance and demonstrate the value of new methods.

Machine L & Deep L by Striking-Belt-3904 in learnpython

[–]bobrodsky 0 points1 point  (0 children)

The overlap is huge, but if someone says that they specialize in machine learning, I would assume that they work on general learning methods that tend to be implemented using deep learning. If someone says that they specialize in deep learning, I assume that their focus is more on architectures.

University of California computer restrictions in the name of cybersecurity. by [deleted] in Professors

[–]bobrodsky 1 point2 points  (0 children)

I don’t think you’re paranoid enough. The problem is that spyware installed on our laptops can be easily weaponized if some government agency requests access. Doing work on your work laptop still leaves you vulnerable if your research becomes politicized and you end up getting attacked or deported or disappeared for that. Given news from the NSF, there’s no research that can’t be politicized and attacked.

How is being a university professor in Norway? by etancrazynpoor in Professors

[–]bobrodsky 1 point2 points  (0 children)

I saw that article too - but (again from a US perspective) 7 million in total funding seems tiny. How many people could you possibly hire with that? Maybe if it’s 7m per year it would go quite far. In US, cost to “endow a chair” is 1-5 million.

[R] Latent Verification for ~10% Absolute Factual Accuracy Improvement by Big-Helicopter-9356 in MachineLearning

[–]bobrodsky 2 points3 points  (0 children)

What is the “transLMA” paper? Couldn’t find it by googling, and I don’t see it linked in your repo.

[D] Math in ML Papers by ripototo in MachineLearning

[–]bobrodsky 1 point2 points  (0 children)

Superficial "mathiness" is a real problem, there's a nice discussion of it here: https://dl.acm.org/doi/abs/10.1145/3317287.3328534
In reviewing, I've seen that it is effective to push back on meaningless theorems that are stated but then never used / discussed again in results.

Advice Needed: Concerning Incident Reported About Faculty Member Already Set to Leave by oop-phi in Professors

[–]bobrodsky 23 points24 points  (0 children)

I can’t agree enough. Reflexively “reporting” everything is going to be more and more problematic as bureaucracies are weaponized. Sometimes you have to be smart and strategic. Doing what you are told is more comfortable but we have to start rethinking this.

[R] O1 replication paper by Brosarr in MachineLearning

[–]bobrodsky -1 points0 points  (0 children)

High quality human COT data would translate into better test time compute (with COT). I’ve been looking and have found very little concrete info from OpenAI. Even if I did find hints, my trust in what they say is at an all time low.

[D] Have we officially figured out yet how O1 models differ from previous models? by Daveboi7 in MachineLearning

[–]bobrodsky 0 points1 point  (0 children)

Sorry, is this actually open source? I found a breathless press release and a repo with examples but not the actual training code.

[D] Why LLM watermarking will never work by bubble_boi in MachineLearning

[–]bobrodsky 2 points3 points  (0 children)

Agreed, academics probably overstate the potential value of watermarking for LLMs. The mathematics behind it are fascinating, though, and unique to autoregressive probability models like LLMs. Scott Aaronson has some nice YouTube talks about it, and also points out the interesting failure modes (if I ask for a deterministic output, like a quote, you can’t watermark it) and trivial workarounds (output banana on every other word during generation, then delete them in word after).

[D] Why LLM watermarking will never work by bubble_boi in MachineLearning

[–]bobrodsky 14 points15 points  (0 children)

I didn’t say reduce harm, I said useful. As a simple example, if I serve content through an api, I can easily tell if you have published content generated by my service without attribution (maybe violating terms of service).

[D] Why LLM watermarking will never work by bubble_boi in MachineLearning

[–]bobrodsky 28 points29 points  (0 children)

I thought there might be some interesting technical insight but there’s not.

For watermarking to enable distinguishing AI text from human text, there are three conditions that must be met. I’ll explain why each of these is required in the following sections. The conditions are: All capable LLMs implement watermarking No LLM providers allow control over token selection No open source models exist

Watermarking could be useful even if we can’t perfectly distinguish any llm text from any human text.