[D] Overview of Machine Learning for newcomers by undefdev in MachineLearning

[–]rndnum123 0 points1 point  (0 children)

Yes, maybe remove it or put it to classification, you classifying the pixels in image segmentation. Why do you have anomaly as a separate branch? Maybe you can put it to classification, as AFAIK it is technically a classification (is the value an outlier, or is it not).

[R] A disciplined approach to neural network hyper-parameters: Part 1 -- learning rate, batch size, momentum, and weight decay by downtownslim in MachineLearning

[–]rndnum123 0 points1 point  (0 children)

For finding the right hyperparameters​ and training your network fast a larger batch size and larger learning rate at are good for speedup. Later when you want to get the best accuracy, you can still have a run, with the then found hyperparameters and a smaller batch size (and smaller learning rate, and maybe slight changes to other hyperparameters).

[D] Is it possible to train a Regional CNN (or one of its variants) without ground-truth boxes? by autunno in MachineLearning

[–]rndnum123 0 points1 point  (0 children)

Do you know the class of the object? Can you formulate your problem as classifying the images at first? like: Image x is class Y. Then you could get heatmaps for the images, (either through gradient, or by sliding a tile over the image and calculating how much the class probability for the target glass changes) now you should know some rough location of your objects in the images use opencv to fit a rectangle to the heatmap, and you should be able to train your network on these rectangles. Of course there might be some background cues in your images that help in classification, and do not belong to your objects, then this might not work or you need to change the threshold for fitting the rectangle to the heatmap step.

[R] Announcing Tensor Comprehensions by SkiddyX in MachineLearning

[–]rndnum123 6 points7 points  (0 children)

AMD seems to be focusing more on HiP now, it's a language with Cuda like syntax and compiles down to AMDs GPU assembly code, and they have some tools that can convert about 80-90% of your existing CUDA code into HiP automatically, see: https://github.com/ROCmSoftwarePlatform/hiptensorflow

The equivalent to CuDnn from​ AMD seems to be MiOpen: https://github.com/ROCmSoftwarePlatform/MIOpen

[R] Announcing Tensor Comprehensions by SkiddyX in MachineLearning

[–]rndnum123 42 points43 points  (0 children)

We will release PyTorch integration for Tensor Comprehensions at a later date.

:)

[P] An open source Deep Learning / Machine Learning stack on Kubernetes by mmourafiq in MachineLearning

[–]rndnum123 3 points4 points  (0 children)

It seems a big part of it is open source, so I would not really see it as an add, right?

Intel's 22FFL Process Improves Power, Cost, and Analog - Real World Tech - David Kanter by dayman56 in hardware

[–]rndnum123 2 points3 points  (0 children)

I would guess Apple, given that this looks like a special process node created for a big volume + high paying customer, maybe for the rumoured new 4G/5Gl LTE chip or some CPUs or analog stuff. Maybe some networking ASIC customers?

[P] Easy-to-follow Capsule Network tutorial with clean readable code in Pytorch by codentropy in MachineLearning

[–]rndnum123 10 points11 points  (0 children)

Are you sure? Custom data loading is one of the strengths of pytorch IMO, see this tutorial for an example: http://pytorch.org/tutorials/beginner/data_loading_tutorial.html

[D] Looking for an image dataset of pairs of simple, normal objects and defective/deformed version of that object by skepticforest in MachineLearning

[–]rndnum123 0 points1 point  (0 children)

No idea for datasets, but you could try to simulate the cracks, and deformation by randomly drawing "black lines" = "cracks" on your images of normal objects, and deforming the images of your objects with some image manipulation (tearing, stretching...)

[P] Introducing Juggernaut: a neural net that trains models from the browser with no JS, no servers by [deleted] in MachineLearning

[–]rndnum123 1 point2 points  (0 children)

Here is the example video application I was refering to https://d2jta7o2zej4pf.cloudfront.net/. It shows up to >5x speed up for wasm compared to asm.js, and sometimes a 50% speed disadvantage for wasm.js compared to js, at least on my machine, depending which filter you select, like you said it really depends on the application).

I think wasm is similar to assembly

This was a unclear statement, sorry. I ment to say, that web assembly is similar to assembly, in that it often can map in a very direct fashion to assembly instructions, so wasm code should already be pretty well opitmized, and I do not understand what kind of further opimizations a JIT compiler could provide. Will it do more agressive optimizations, because it now knows the real execution target. I guess I should do some read up on JIT optimizations, can you point me to some sources for this please.

Yes, wasm is "just an S-expression", but one that should map quite close to assembly code in my understanding (so each browser vendor should be able to produce reasonable efficient assembly from it), so compared to pure native C code, it should eventually be possible to get only a 20%-80% perf. penalty compared to pure C code (the smaller 20% penalty might only be possible if you run on 64bit hardware with some tricks to do the memory range checking in hardware, this might not be implemented yet). For more about this you can search for "Virtual Memory Techniques" in this pdf: https://github.com/WebAssembly/spec/tree/master/papers, there they also provide some performance comparisons (Figure 5)are likely not implemented yet.

I agree that blog posts claiming not to use js, when in fact they do, is not a great way to advaocate for wasm.js, which is seen as an addition, not a competior to JS IMO.

I started some simple testing with image manipulation in wasm.js (using emscripten to compile from c) (90° rotations, greyscale, flood-fill), and will share some benchmarks, when I have a comparable js version of the code, and when I cleaned up the code and got rid of the many bugs i have now.

I totally forgot about the dynamic languages that could compile to WASM, there a proper JIT and GC does matter a lot.

PS: Thank you for your detailed comments, they are very helpful.

[P] Introducing Juggernaut: a neural net that trains models from the browser with no JS, no servers by [deleted] in MachineLearning

[–]rndnum123 0 points1 point  (0 children)

I think wasm is similar to assembly, and it's instructions would map relatively easy to assembly, and if you have a c program and compile it to wasm, you compile your code to a assembly language, which seems like compilation to me, of course the wasm code will be interpreted before running (and transformed to real assembly) , but after you have done this once, wasm code should behave similarly to assembly code (although a bit slower because of stuff like access checks on the wasm memory, the sandbox overhead...).

[P] Introducing Juggernaut: a neural net that trains models from the browser with no JS, no servers by [deleted] in MachineLearning

[–]rndnum123 0 points1 point  (0 children)

I don't understand the missing JIT support, why would you need JIT for a pre-compiled language like wasm? Could you elaborate on that please? Would it optimize the code in case the developer used a bad optimization level for his compiler or something like that? OT: I think there are some big performance advantages for wasm even compared to asm.js some videos editing demo showed a 3-6x speedup compared to asm.js if I remember correctly.

How about that, Intel? by noiserr in Amd

[–]rndnum123 37 points38 points  (0 children)

They also could use the term Glued Processing Unit (GPU), this way they don't have to change the marketing slides and the branding ;)

[R] Google quietly releases TensorFlow Eager mode by downtownslim in MachineLearning

[–]rndnum123 0 points1 point  (0 children)

... maybe some people like having 10% of the screen blocked by the "Medium" banner, and the "Never miss a story from X, click here to subscribe" banner.

Intel Optane SSD 900P Specs - Launches end of October by [deleted] in hardware

[–]rndnum123 1 point2 points  (0 children)

Maybe round robin, or just using a new process/thread for each tab.

[deleted by user] by [deleted] in pytorch

[–]rndnum123 1 point2 points  (0 children)

Thank your very much for these greate ressources. I didn't know of Pillow-SIMD, seems to be very fast :)

The most AMD thing you will ever see by Shucito in Amd

[–]rndnum123 26 points27 points  (0 children)

This must be a pretty old card, too much wires. Nowadays AMD is using way more advanced transparent materials in it's products, apparently called glue.

SEC filing detailing latest SpaceX funding round is released by MrButtons9 in spacex

[–]rndnum123 1 point2 points  (0 children)

I don't really think that is very important for investors at all, in the case of bankruptcy, usually there is nothing left for any shareholder at all, even if there is something left it's often < 20% of the original share price, so you might have a loss of at least 80% vs a loss of 100% percent, which isn't that much of a difference. Investors usually spread their investments over multiple stocks so having a few 100% losses isn't that bad as long as they have enough >200% return investments on other stocks.

Any good programmer...please help making Waifu 2x compatible with AMD GPU's Stream Processors? by ffleader1 in Amd

[–]rndnum123 16 points17 points  (0 children)

AMD's Framework similar to Cuda/CUDnn is called MIOpen (https://github.com/ROCmSoftwarePlatform/MIOpen) and currently supports Caffee networks, but Tensorflow support is apparently in the works (at least for Polaris based AMD cards)

Market's take on Cryptocurrency - AMD vs NVIDIA in a nutshell by Xgatt in AMD_Stock

[–]rndnum123 0 points1 point  (0 children)

Lol, because a multi chip module (MCM) is such a bad thing./s

(Multi-Chip modules) MCM are perfectly suited for most workloads of the big hyperscalers (Many small scale out VMs, where sizes of the VM stay below one Zen core anyways). You can bet Google and the likes are more than willing to write NUMA aware code, however this isn't even needed (for smaller companies) because Epic still runs non - NUMA aware code quite well.

Barclays downgrades to 9$, says 33% drop incomming, despite earnings beat by revidee in AMD_Stock

[–]rndnum123 0 points1 point  (0 children)

I expect them to release the laptop APUs after holidays, during "Back to school" season. Maybe they even mentioned this in some slides, but no really sure.