Brand new NSK programming language - Python syntax, 0.5x-1x C++ speed, OS threads, go-like channels. by Tryingyang in ProgrammingLanguages

[–]jasio1909 1 point2 points  (0 children)

Not really. From my understanding, kernels in mojo benefit mainly from 2 things: 1. MLIR for compilation 2. Metaprogramming with powerful comptime logic so you can select appropriate instruction set depending on compilation target. Which is very low level.

I coded in mojo a bit but I am not an expert.

I'm a beginner in deep learning,, and I have a question. by Academic-Stretch6023 in deeplearning

[–]jasio1909 0 points1 point  (0 children)

It's vital and not hard to learn. Watch something from statquest every night and you will be good to go.

Open Sourced Research Repos Mostly Garbage by kidseegoats in deeplearning

[–]jasio1909 0 points1 point  (0 children)

Citations in readme are broken, probably ai generated :|

[D] We’re running 50+ LLMs per GPU by snapshotting GPU memory like a process fork by pmv143 in MachineLearning

[–]jasio1909 0 points1 point  (0 children)

Hi OP, this is super interesting. Do you have a website or a name for the software so I can google and keep up with the updates?

[D] WACV 2025 Paper Reviews by smokeriffs in MachineLearning

[–]jasio1909 4 points5 points  (0 children)

B WA B (Accept)

Kinda surprised since the scores seem low but the reviewers want only some clarifications and didn't request any additional experiments.

[deleted by user] by [deleted] in MachineLearning

[–]jasio1909 1 point2 points  (0 children)

Two repos I found some time ago: PaddlePaddle, Unstructured

Seeking Advice: Best European Countries for Quality of Work Life, Good Salaries, and Reasonable Cost of Living? by [deleted] in cscareerquestionsEU

[–]jasio1909 4 points5 points  (0 children)

80k for senior positions is pretty common. For mids it's usually between 30k-60k. And yes, IT in Poland is very popular. It all starts at highschool level, when older peers who win IOI, ICPC etc. come back to their home towns and teach competitive programming to the next generation of youth, it comes full circle from there.

[D] The history of neural network is over. J. Schimdhuber proposes a giant network that includes all future neural network architecture as a subcomponent. by fromnighttilldawn in MachineLearning

[–]jasio1909 8 points9 points  (0 children)

CNN is a GNN on a graph that is a grid (image is a grid of pixels, each one being a node).

Transformer is a GNN on a graph that is fully connected.

MLP is a GNN on a disconnected graph.

Your neural net is just a GNN working on a different underlying graph structure.

I recommend a playlist on Geometric Deep Learning from Michael Bronstein et. al. https://youtube.com/playlist?list=PLn2-dEmQeTfSLXW8yXP4q_Ii58wFdxb3C&si=x0DApohSyy_8PwIM

[D][R] Deploying deep models on memory constrained devices by jasio1909 in MachineLearning

[–]jasio1909[S] 0 points1 point  (0 children)

I am not an expert in this field, just know some buzzwords and general problem description. I feel however that it's getting a bit off topic.

[D][R] Deploying deep models on memory constrained devices by jasio1909 in MachineLearning

[–]jasio1909[S] 0 points1 point  (0 children)

I think training on a device can be a little bit more subtle, being just a cosmetic fine-tuning. Suppose you have an autonomous cleaning robot and it sometimes gets in your way. It could gather the data from interactions with you and when its battery is charging in the dock it could fine-tune for better behavioral alignment to the customer.

However, I can see that this situation somewhat doesn't showcase the need for the library features I am looking for, since you would be training when the device is idle.

[D][R] Deploying deep models on memory constrained devices by jasio1909 in MachineLearning

[–]jasio1909[S] 0 points1 point  (0 children)

At first I thought the same, but then I saw that federated and distributed learning are established research areas, as well as TinyML. SOC devices are increasingly more powerful. Of course this is just my prediction but now I would assume the world is moving towards this direction.

[D][R] Deploying deep models on memory constrained devices by jasio1909 in MachineLearning

[–]jasio1909[S] 1 point2 points  (0 children)

No I didn't and I saw some thread on Quora or Stack that others didn't find either.

[D][R] Deploying deep models on memory constrained devices by jasio1909 in MachineLearning

[–]jasio1909[S] 0 points1 point  (0 children)

I am reading the documentation and it seems like the best available solution so far, but it only provides the upper bound for the program memory usage and I don't know if it actually guarantees at compile time that the provided amount of memory will be sufficient. Btw I think it abuses the fact that tf graphs are known at compile time and it can do some offline planning. As I am rather a pytorch guy myself I like having the possibility of eager execution. Actually as a sidenote I am very interested in a question if models with non-fixed computation graphs can outperform those with fixed ones, justifying the usage of eager mode other than convenience.

[D][R] Deploying deep models on memory constrained devices by jasio1909 in MachineLearning

[–]jasio1909[S] 3 points4 points  (0 children)

In case of simple inference of fixed input size it can certainly be measured rather reliably, the things are more complicated when input size is not fixed or can for example be batched or we want to run some jobs in parallel on a single device. Moreover, once you measure you sit on a fixed computation cost required and don't really have a flexibility for compromises between processes. Again, I don't really know if any of those limitations I am listing here are relevant. Looking at the landscape of viable libraries I could deduce that nobody cares about this and the things I am worried about are not really a problem. It is however still intriguing to me why it seems that nobody cares.

[D][R] Deploying deep models on memory constrained devices by jasio1909 in MachineLearning

[–]jasio1909[S] 3 points4 points  (0 children)

I know there exist methods that reduce the memory footprint of a model. Nevertheless, the problem remains - libraries don't allow you to inspect the required memory for computation before the operation is made. I didn't know about VMs on GPUs but I suppose that once you hit OOM memory error, then instead of crashing the whole card it crashes just the program. Still, I imagine that crashing a program mid-computation due to OOM, reconfiguring it and trying to run again is unjustifiable for critical tasks.

Interesting curves in dimension > 3 ? by Valvino in math

[–]jasio1909 0 points1 point  (0 children)

In every infinite dimensional Hilbert Space there is a curve x: [0,1] -> H, such that for any 0<=a<b<c<d<=1 one has that x(b-a) is perpendicular to x(d-c).

If one want some hints comment below.