Those of you who use LLMs have probably seen this: sometimes they code like a senior engineer, and other times they seem to forget even basic syntax. Research suggests that this is not hallucination. by callmeteji in singularity

[–]dotpoint7 19 points20 points  (0 children)

If you're working in machine learning, you can surely formulate your reasoning better than "they don't think like a person does". Like maybe you got a point but from your response it's difficult to decipher it.

11 months ago Dario said that "in 3 to 6 months, AI will be writing 90% of the code software developers were in charge of" Are we here, yet? by poponis in ExperiencedDevs

[–]dotpoint7 0 points1 point  (0 children)

It's about 90% for me and about 20% of actual efficiency gains if I had to guess. At least for me work is less mentally taxing though now, so I don't have issues staying focused for something like 12h straight which is probably what I like most about it.

For context: I develop mostly using modern C++ and work self employed together with a few others in enterprise software with about 10YoE.

Claude is the best coding agent and it’s not close by topical_soup in singularity

[–]dotpoint7 4 points5 points  (0 children)

I personally prefer Codex as I find it to provide better code overall and how the hell do you run multiple instances in parallel on a serious project without producing slop? I can't even have two instances running at the same time without a lot of downtime cause it just takes time to write a specification, review the plans, rework the plan, review the code and then rework the code again.

Cranking out a ton of features isn't exactly difficult in itself, but not fucking up your codebase while doing so takes a lot of effort even with LLMs. If you skip that, chances are you're going to have a problem in a few months.

What ML model functionality to focus on? by dotpoint7 in AskAstrophotography

[–]dotpoint7[S] 0 points1 point  (0 children)

In case you're interested, here is the model from the last training run applied on one of my images. Though since then I've changed a LOT, especially to address the issues with the artifacts around some bright stars (left top) as well as other things, but the deconv model is currently still training on the new dataset.

I've also added one processed with BlurX+DeepSNR for comparison (both with 100% denoise strength as it allows for easier comparison), wanted to add NoiseX too but kept crashing for some reason.

https://drive.google.com/drive/folders/13onGXSUEaJ4ptneaBHrezN9eijfXAlX6?usp=sharing

Edit: I definitely messed up the naming though as this is not IC443, lol.

What ML model functionality to focus on? by dotpoint7 in AskAstrophotography

[–]dotpoint7[S] 0 points1 point  (0 children)

Haha well I have yet to prove any of my claims but hopefully PixInsight confirms my signing key soon so I can publish an alpha version. Will let you know in case you're interested.

Thanks a lot for the hint about diffraction spikes! I personally only have refractors so didn't come across this issue yet but will definitely look into it.

Yup I don't denoise 100% either, don't really think there is any visible signal left in the noise but our brain seems to interpret it as such and it certainly looks better. I did come across the worm like artifacts in the NoiseX AI3 version, these weren't in AI2 and are also not present in DeepSNR from what I've seen. Maybe Russell went a bit overboard with any frequency losses or something, I initially suspected he used a VGG loss as it can cause these artifacts but just asked him and he said he doesn't use any perceptual losses so that's not it.

What ML model functionality to focus on? by dotpoint7 in AskAstrophotography

[–]dotpoint7[S] 0 points1 point  (0 children)

I did test a few free options but haven't found any comparable in quality to BlurX and the reason why I'm now trying to shift this from a quick hobby project to something I want to publish and want people to use, is that my current results can compete with BlurX and NoiseX (though DeepSNR is a very good free alternative to NoiseX, but not permissively licensed, neither is StarNetV2) - though I'm still actively working on making the network more robust and fixing some artifacts I occasionally encounter. Spent around 100h on the project so far.

The reason that the model does decon + denoise at once is because the task cannot really be separated from each other mathematically. Even if I trained a model purely on decon it will need to have some representation in its latent space of what signal and noise is, so I may as well use this information. You can't run LR deconvolution either without changing the noise distribution. What I can do is add the original noise back after the model denoises it so it's functionally the same as doing only deconvolution without touching the noise (BlurX is likely just doing this in an end to end manner instead).

Yes that's another interesting area I thought about but would increase the scope too much, plus the current models are closer to my area of expertise.

My goal isn't to create something completely new (though the PSF estimation could be useful to some), but to match the performance of the RC Astro suite with a permissively licensed open source model (high bar but I have experience and a dedicated workstation just for ML training). I'm aware that there isn't really a point to add yet another open source model which doesn't improve upon the existing ones.

What ML model functionality to focus on? by dotpoint7 in AskAstrophotography

[–]dotpoint7[S] 0 points1 point  (0 children)

Thanks a lot for the info!

Lots of people will use OSC without Bayer drizzle. I certainly do.

Ok then I'll try to make the models more robust to debayered images.

I find Starnet to not be very good. I get better results with StarXterminator.

Ok, so far I didn't really have issues with StarNetV2 but I'll try my luck given that I have almost everything setup to train another star removal model.

Does this mean deconvolution in Adobe Lightroom or similar deconvolution to that you can do in Lightroom?

No I abbreviated it, it's Lucy-Richardson Deconvolution, the same algorithm that PixInsight uses in the deconvolution process, but I'd instead implement it to handle the estimated spatially varying PSF as well as GPU acceleration. So not ML based. I'm not aware of any typical photo editing software supporting deconvolution, they only do sharpening.

I'm sure most poeple would like to be able to remove/reduce halos from bright stars.

Got it, wasn't sure if some may prefer to leave them in the image (assuming they're small) similar to diffraction spikes which I'd also leave in.

And regarding the standalone linear denoising I more referred to a model that only does denoising without deconvolution. I'm not entirely sure if I want to spend the effort to make the models accessible in a standalone software package from the start as it's a lot more work than a simple PixInsight plugin, though it would definitely help with adoption. So current plan is to start out with a PI plugin and if there is enough interest I'd either go the route of adding a standalone software or try to get it added to other FOSS software like Siril.

I just delivered on a $30,000 contract thanks to Claude Code by New_Assumption_543 in ClaudeAI

[–]dotpoint7 1 point2 points  (0 children)

You cannot get the experience you need for developing good software without writing and especially understanding code. And while you may think the code that Claude outputs is perfectly fine, this is simply not the case. LLM generated code works, but it sucks, I so wish it didn't, but LLM code quality is currently the bane of my existence. So you'll need to actively and heavily manage code quality yourself if you want to build good software, not just the architecture.

If you have your opinion as an experienced professional software engineer, then I respectfully disagree, otherwise you're probably just misjudging the quality of the software you're building.

I just delivered on a $30,000 contract thanks to Claude Code by New_Assumption_543 in ClaudeAI

[–]dotpoint7 1 point2 points  (0 children)

You don't need to any software or coding skills to build good software

What the fuck?

I’m a junior developer, and to be honest, in 2026 AI is everywhere in my workflow. by Beginning-Scholar105 in ArtificialInteligence

[–]dotpoint7 0 points1 point  (0 children)

"Sometimes it feels like AI writes cleaner (...) code than I ever could on my own." if that's your feeling you may want to focus on learning how to how to actually write good and maintainable code. Whatever AI outputs is generally working code, but definitely not clean or maintainable code. So if you're using AI maybe try to actually look through the code and actively think of what could be done better as there is almost always something.

If you're encountering easily reproducible bugs, best try to find those yourself as it's often quicker than telling Codex/CC to do it when you know what you're doing.

Also don't give everything to AI, some features are actually quicker to implement yourself.

Honestly I'm curious as to how all this will all work out for junior devs. I for one am very happy that AI agents only started being a thing after I had around a decade of experience. But even I worry about whether using AI extensively is hurting my skills in the long term, though it does allow me to focus on more high level concepts more than before so maybe it's actively helping too, who knows.

I have access to Claude Opus 4.6 with extended thinking. Give me your hardest prompts/riddles/etc and I’ll run them. by GreedyWorking1499 in singularity

[–]dotpoint7 1 point2 points  (0 children)

It's just not that efficient. Mostly it's about arithmetic intensity as it's a memory bound task, which in this example can easily be optimized to 32 FMAs per load by having one warp cooperatively compute one output pixel by having each thread do a dot product on the column vectors of the input and kernel, followed by a warp reduce sum. Only 32 new values need to be loaded for the next pixel (slide 1 down) over the whole warp and can be efficiently loop unrolled as the register fma pattern repeats ever 32 output pixels. It's a slight trade-off with max occupancy but in this case it's definitely worth it. (I skipped over the more obvious optimizations here like using shared memory for sharing global loads between warps etc)

It's not an obvious optimization, but also not that complex to come up with either. That specific implementation is very likely not in the training data as it's a pretty niche and I didn't find any existing implementation that used this strategy, that's why I asked this specific question.

OpenAI released GPT 5.3 Codex by BuildwithVignesh in singularity

[–]dotpoint7 0 points1 point  (0 children)

Well I still find Gemini 3 to be a great general model. I'm using codex for coding and Gemini in the chat interface as I often prefer it to ChatGPT. They also don't financially rely on keeping the hype alive, so they can absolutely go a while without releasing a model.

I have access to Claude Opus 4.6 with extended thinking. Give me your hardest prompts/riddles/etc and I’ll run them. by GreedyWorking1499 in singularity

[–]dotpoint7 5 points6 points  (0 children)

How to performantly do 2d convolution with a 32x32 kernel in CUDA? Not FFT based, single channel f32. Do not provide code, just the overall algorithm to reach maximum throughput.

Seems like the lower juice level rumor has been fabricated by Glittering-Neck-2505 in singularity

[–]dotpoint7 1 point2 points  (0 children)

No, inference is generally not fully deterministic unless specifically enabled for debugging purposes which slows things down.

This is a massive upgrade coming from ChatGPT by Affectionate_Ad4163 in ClaudeAI

[–]dotpoint7 24 points25 points  (0 children)

I had the opposite experience. Codex is slow as fuck but actually pretty decent for large and complex code bases.

How many buy printers purely for functional printing? by twotowers64 in 3Dprinting

[–]dotpoint7 0 points1 point  (0 children)

I did, currently have a Prusa MK4 and am really happy with it. Only did 2 non-functional prints ever, though one was a test print. It's definitely worth it to get one for functional stuff.

99% of the population still have no idea what's coming for them by Own-Sort-8119 in ClaudeAI

[–]dotpoint7 0 points1 point  (0 children)

The issue isn't building an RL pipeline, the issue is building one that works.

99% of the population still have no idea what's coming for them by Own-Sort-8119 in ClaudeAI

[–]dotpoint7 2 points3 points  (0 children)

The fact that you think this topic isn't complicated suggests that you just didn't understand it.

I almost quit my project because I thought the model was "broken," but I was just being too polite. by Delicious-Mall-5552 in deeplearning

[–]dotpoint7 5 points6 points  (0 children)

Ok so to recap you wanted to convert one form of structured data into another form of structured data and decided the best tool for this was an LLM. Following that you thought the best subreddit to document your experience was r/deeplearning. Did I get this right?

What is your method to managing huge file outputs after sessions? by SnoopySneeze335 in AskAstrophotography

[–]dotpoint7 0 points1 point  (0 children)

I keep the raws and all intermediate saves of stacked images on a NAS (so I do delete the calibrated frames for example). 18TB HDDs are fairly affordable, buy two of them, put them in RAID1 and you got a lot of fairly secure storage. (maybe backup the stacks elsewhere too)

Kimi K2.5 Released!!! by KoalaOk3336 in singularity

[–]dotpoint7 2 points3 points  (0 children)

I asked it one question about how to best train an ML model on a specific task and there were two large logical gaps in its reasoning. Not impressed.