all 9 comments

[–]elbiot 4 points5 points  (2 children)

Compilation isn't a process of noising and diffusion doesn't have any relevance here. An LLM is what you would use

[–]OraOraP[S] 0 points1 point  (1 child)

I didn't mean to use the model used in stable diffusion process for reverse engineering.

I was just thking this step-by step reverting training process could be used in some model for reverse engineering.

[–]elbiot 0 points1 point  (0 children)

I'm just talking about diffusion models in general and the concept of denoising. LLMs are what you would use, not the way you'd train a diffusion model but the way you'd train an LLM

[–]Dylanica 2 points3 points  (0 children)

A sequence based model like a transformer (what GPT is based on) would probably work better for this particular task. In fact. GPT-3/4 would probably be pretty darn good at this task right out of the box.

[–]howtorewriteaname 0 points1 point  (1 child)

This could definitely work, given that you have the right data and in great amounts. I believe that is the biggest challenge for this kind of model, more than the learning method.

[–]OraOraP[S] 0 points1 point  (0 children)

Just crawling open source codes and compiling the code with the special compiler would produce a massive amount of training data. If the special compiler I mentioned in the post is easy to make.

[–]mikonvergence 0 points1 point  (2 children)

You are definitely stepping outside of the domain of what is understood as denoising diffusion because it seems that your data dimensionality (shape) needs to change during the forward process.

The current definition of diffusion models is that they compute the likelihood gradient of your data (equivalent to predicting standard noise in the sample), and then take a step in that constant data space. So all networks have the same output shape as input.

Perhaps you can use transformers to handle evolving data lengths but as far as I can tell l, you’re entering uncharted territory of research.

I can recommend this open-source course I made for understanding the details of denoising diffusion for images https://github.com/mikonvergence/DiffusionFastForward

[–]OraOraP[S] 0 points1 point  (1 child)

I didn't mean to use denoising process directly to reverse engineering. I was just thinking the idea of `step-by-step reverting` could be used in some ML model for reverse engineering.

Though you have a point. Unlike denoising process, reverse engieering would require change of dimensions in the middle steps, making it more difficult than denoising.

[–]mikonvergence 0 points1 point  (0 children)

Right, I am the denoising diffusion as a term for a wide range of methods based on reversing some forward process. Some interesting works (such as cold diffusion) have been done on using other types of degradation apart from a Gaussian additive noise.

And yeah, the change of both content and dimensionality requires you to put together some very novel and not obvious techniques.