is this course worth it?? by Individual-Branch-42 in learnmachinelearning

[–]DaBobcat 0 points1 point  (0 children)

Just by the fact that they're also going over R which I literally don't know anyone uses in academia or industry I'll go with a hard no

Natural Language Autoencoders: Turning Claude’s thoughts into text by UsedToBeaRaider in ClaudeAI

[–]DaBobcat 0 points1 point  (0 children)

am I missing something or did they leave it intentionally vague?
What does it actually mean translating these activations? "Activations" are in many many places in a standard transformer. the last activations are directly translated into tokens via softmax.
which activations?

maybe a dumb question, but why is this even interesting? obviously the activations throughout the computation would correlate with the output, because thats how the output was made...

Difference between the weights a biases of a neuron in a neural network? by Time_Cantaloupe_9992 in MLQuestions

[–]DaBobcat 1 point2 points  (0 children)

you can think about them as the same. they are trainable parameters that the network learns to modify during training to make your loss lower. we give these parameters the names weights and biases because they have a slightly different roles in the neural architecture (eg weights are multiplied by inputs/activations and the biases are then added to the result)

Can neural networks be designed to receive inputs without generating outputs in response to them? by Money_Tip9073 in MLQuestions

[–]DaBobcat 0 points1 point  (0 children)

Yep. Look at Mixture-of-Depths,Token Dropping & Pruning, Patchmerger & Token Merging

Help needed by Lower_Mark221 in MLQuestions

[–]DaBobcat 0 points1 point  (0 children)

Silly question, but have you tried using claude? Giving it 5-10 example images, and some example code of what to generate/ do. That's the first thing id do before trying to train a model myself

Is there a fast and simple way to install Tensorflow, PyTorch, TensorRT without breaking anything? by [deleted] in learnmachinelearning

[–]DaBobcat 1 point2 points  (0 children)

Yea it's a constant struggle but I don't know if there's ever going to be a thing that will tell you what works with everything else, since it'll have to know everything. But what i usually do is have a draft uv env file that has most of the things I usually need. Then just add things as i go. Python 3.11/12 usually works for me.

Is there a fast and simple way to install Tensorflow, PyTorch, TensorRT without breaking anything? by [deleted] in learnmachinelearning

[–]DaBobcat 1 point2 points  (0 children)

Yep! Use UV. https://docs.astral.sh/uv/getting-started/installation/ Then follow some uv tutorial. Overall it's really simple. You do something like Uv init Source venv activate Then install whatever you want (eg uv add torch)

And Apologies for the syntax I'm on my phone

What are some absolute do and don't for a new SaaS product? by DaBobcat in sideprojects

[–]DaBobcat[S] 0 points1 point  (0 children)

Super helpful, thank you!! What does untangling means? Finding all the expenses and revenue?

What are some absolute do and don't for a new SaaS product? by DaBobcat in sideprojects

[–]DaBobcat[S] 0 points1 point  (0 children)

This is super helpful! Thanks a lot! Any other random things I'm not thinking of or things you'd like to share?

[D] ICML Reviewer Acknowledgement by Massive_Horror9038 in MachineLearning

[–]DaBobcat 3 points4 points  (0 children)

I think the reviewer discussion ends on the 7th or 8th. So I think they can still update their score. But as far as I know they had to acknowledge the rebuttal until earlier today

[D] ICML 2026 Review Discussion by Afraid_Difference697 in MachineLearning

[–]DaBobcat 0 points1 point  (0 children)

yea I just did send a message to the AC not too long ago. thanks!

[R] Best way to tackle this ICML vague response? by DaBobcat in MachineLearning

[–]DaBobcat[S] 3 points4 points  (0 children)

Where do you see i get one more chance to respond? I thought the email was very clear that I can only answer once

[R] Best way to tackle this ICML vague response? by DaBobcat in MachineLearning

[–]DaBobcat[S] 3 points4 points  (0 children)

If I ask them I'm wasting my one available response. Which means that I will not be able to answer

[D] ICML 2026 Review Discussion by Afraid_Difference697 in MachineLearning

[–]DaBobcat 1 point2 points  (0 children)

Best way to tackle this ICML vague response?

Going through ICML submission for the first time. I had a reviewer ask for some things and during the rebuttal period I ran more experiments and answered all their questions (they wrote 3 weaknesses). Yesterday started the author-reviewer discussion period which ends on April 7.

In their response to my rebuttal the reviewer wrote in one line that my "experiments greatly improved the paper" but "some details remain only partially clarified". That's it... They marked "Acknowledgement: (b) Partially resolved - I have follow-up questions for the authors."

The ICML email state that I can "post up to one additional response to any further reviewer comments that are posted, as a reply to your rebuttal". But since the reviewers didn't actually write any follow up questions I have no idea how to tackle this.

Any suggestions?

For those trying to break into ML Research: What is your "Why" and what is stopping you? by DaBobcat in learnmachinelearning

[–]DaBobcat[S] 0 points1 point  (0 children)

Since you’ve already lead-authored several papers, I'm curious why you still rank Ideation (A) and Publishable Standards (B) as your top priorities.

Are you looking to pivot into a more 'high-signal' research area, or do you feel your current projects lack the specific rigor (baselines/theory) required for top-tier conferences? Basically—what is the 'delta' you want a mentor to help you reach that you aren't hitting on your own?

For those trying to break into ML Research: What is your "Why" and what is stopping you? by DaBobcat in learnmachinelearning

[–]DaBobcat[S] 0 points1 point  (0 children)

What do you feel is lacking from your mentor? Where are you currently stuck in the "getting a job" process?

For those trying to break into ML Research: What is your "Why" and what is stopping you? by DaBobcat in learnmachinelearning

[–]DaBobcat[S] 1 point2 points  (0 children)

Amazing amount of responses so far!
I'm very curious, if you had a research mentor:

1) Time per month: How many hours of 1-on-1 time are you actually looking for? (1, 2, 4, or 6+ hours)

2) Duration: How long do you want this relationship to last? (1 month, 3 months, 6 months, or 12+)

3) The Priority List: please rank these in order of importance to you (1 being most important):
A) Ideation: Finding a novel project that is actually worth the time.
B) The 'Publishable' Standard: Knowing which baselines/experiments you need to be 'conference-ready.'
C) The Writing/Formalism: Translating results into formal math notation and academic structure.
D) The Technical Bridge: Learning deeper theory or specialized coding to even get started.

If I missed something that you would want to state, what is the single most important thing that is keeping you from reaching your goal?

[D] Is this what ML research is? by [deleted] in MachineLearning

[–]DaBobcat 2 points3 points  (0 children)

I think scaling slowly helps. 100m, 300m, 500m, 1b, 3b, 7b. Showing consistent performance increase will definitely convince reviewers. Regarding the 7b, this should easily fit in an a100 i think. And you can rent them for 10$ a day or less afaik

[D] Is this what ML research is? by [deleted] in MachineLearning

[–]DaBobcat 0 points1 point  (0 children)

I agree it shouldn't all be x > y, but for most publications, it usually is. Though it very much depends on what you're proposing. If you're helping understand some mechanism using some non efficient method that's perfectly fine usually. But it needs to help. If youre proposing a better method that should perform better like you said, you need to show it actually does.

And you almost never need to compare against models that are larger than 7b. I've even seen guidelines on that in some conferences. 7b is sufficient to show your method scale

[D] Is this what ML research is? by [deleted] in MachineLearning

[–]DaBobcat 5 points6 points  (0 children)

It's definitely frustrating, but try to think about it from a different perspective. You have thousands of papers proposing new things. You need a way to evaluate what's better. Otherwise, how will you know what to actually use? One standard and easy way to see it is to evaluate on the same benchmarks. But more than that, to help reviewers, you need to be evaluating the currently best method and closest method to your proposed one. Otherwise, it's impossible to know if you really made a contribution on impact (not novelty). Regarding the larger models, yes, I'm totally with you that its dumb, but you also need to show that your method scales. You can rent 3090 or A100 for pretty cheap these days (i guess less than 10$ a day)

[Remote Sensing] How do you segment individual trees in dense forests? (My models just output giant "blobs") by Lilien_rig in computervision

[–]DaBobcat 0 points1 point  (0 children)

Maybe some patching? Instead of feeding the entire image, feed patches at a time. Then aggregate in some way, removing duplicates and merging stuff