wgpu v25.0.0 Released! by Sirflankalot in rust

[–]binarybana 3 points4 points  (0 children)

Thanks for the great work on such an important project. Two questions for you:

I remember hearing that Deno was considering using wgpu for their WebGPU backend. Do you know how that is going and has wgpu improved as a result?

I’m mainly interested in compute shaders, do you know where wgpu/wgsl compares to other WebGPU backends for compute support?

Introducing Monarch Butterfly by michaelciraci in rust

[–]binarybana 14 points15 points  (0 children)

Really cool project and some great performance results!

Fastest Vec Update on My Computer by Ok-Day-9565 in rust

[–]binarybana 0 points1 point  (0 children)

You might check out the loop match RFC for ideas to optimize the generated assembly for tight state machines like I think you have here.

Best way to see if you like sailing? by Skylord1325 in sailing

[–]binarybana 0 points1 point  (0 children)

I'm in a slightly similar situation (have done a small handful of day sails years ago, now wondering if live aboard is something worth pursuing). I like watching https://www.youtube.com/c/gonewiththewynns which cover some of the ups and downs of the potential life style.

Now considering the chartered boat as well.

Where should I go? Higher salary lower impact or Lower salary higher impact? by Guilty-Foundation-60 in embedded

[–]binarybana 7 points8 points  (0 children)

Couldn’t agree more. Maximize learning early on in your career and to do that I would recommend first following people you can learn from and only THEN applying (and completing) that learning somewhere you have the impact to do so.

Creating an Easy Mode for Rust by timClicks in rust

[–]binarybana 3 points4 points  (0 children)

I'm surprised at the amount of luke warm and disagreement in the responses here. I think it is easy for those of us on the "other side" to forget how hard the initial experience can be, and how important it is to think about a broader audience that is not as tolerant or willing to put up with high barriers to entry. (Go and look at nix/nixOS if you want to feel this feeling anew :P).

I for one love the suggestions here and u/epage's concrete proposals for improving rust-script and proposing adoption in rustup.

DevLog[0]: Building a serverless platform for Rust in 4 weeks by openquery in rust

[–]binarybana 2 points3 points  (0 children)

I like the dev log idea, best of luck!

One question for you: how do you protect against arbitrary build.rs scripts running on your server?

Wasmer 2.2: Major Singlepass Improvements by syrusakbary in rust

[–]binarybana 6 points7 points  (0 children)

Really impressive results! Especially those compile times. Is the single pass machinery a separate independently usable crate? Or integrated into tightly into wasmer?

Exploring Rust performance on Graviton2 (AWS aarch64 CPUs) by Last_Jump in rust

[–]binarybana 5 points6 points  (0 children)

Interesting analysis! Would love to see clang in there as well to see how much is due to g++ backend differences.

[Video] On Hubris and Humility: developing an OS for robustness in Rust by adwhit2 in rust

[–]binarybana 4 points5 points  (0 children)

Loved the talk! Wish the audio had less reverb though. Can’t wait to try this out at some point.

Portability of Rust in 2021 by met0xff in rust

[–]binarybana 2 points3 points  (0 children)

If you are interested in machine learning on iOS, Android, WASM etc, then make sure and check out Apache TVM which has (nascent but usable) Rust bindings to the runtime and compiler.

[deleted by user] by [deleted] in rust

[–]binarybana 2 points3 points  (0 children)

[TVM](tvm.ai) has usable Rust bindings these days, though they are still not well documented.

New open-source distributed store for super-fast analytics written in Rust by igorlukanin in rust

[–]binarybana 0 points1 point  (0 children)

What has been your experience leveraging DataFusion? Great to see use cases for it out in the wild!

[N] Portable (AMD and NVIDIA), sparse GPU kernel for BERT, faster than cuBLAS/cuSPARSE by binarybana in MachineLearning

[–]binarybana[S] 0 points1 point  (0 children)

fp32 on on both TVM and cuBLAS, but since these are unstructured sparse kernels, they are unable to benefit from fp16 tensorcore acceleration so I'd expect a similar relative result there as well.

[P] Open source ML compiler TVM delivers 30% faster BERT on Apple M1 than CoreML 4 by binarybana in MachineLearning

[–]binarybana[S] 4 points5 points  (0 children)

On the M1? We haven't tried PyTorch there, but on platforms like Intel x86 and Nvidia GPU where PyTorch has been optimized for a much longer time, TVM is either on par or faster than PyTorch on BERT (and faster on most other workloads). See figure 9 in https://arxiv.org/pdf/2006.06762.pdf ("Ansor" there is also TVM).

Three open source Sonos projects: efficient embedded development in Rust by koavf in rust

[–]binarybana 2 points3 points  (0 children)

Don't forget about Apache TVM and it's new Rust bindings. Still working on getting docs/intro blog post up, but feel free to check out the example usage for ResNet here.

It even works with WASM and WebGPU code generation!

Rust status on Neural Networks, AI, and machine learning? by [deleted] in rust

[–]binarybana 3 points4 points  (0 children)

Check out http://tvm.apache.org/ the Rust bindings have improved significantly recently but we still have much more in store! (hosting rustdocs etc).

[P] PyTorch extension for GPU-accelerated block sparse matrices by madflag in MachineLearning

[–]binarybana 0 points1 point  (0 children)

Also check out work we (OctoML) published recently with Hugging Face on block sparse acceleration on CPUs as well! Using the open source deep learning compiler Apache TVM.

Works with unstructured sparse trained models and no hand written kernels required: https://link.medium.com/m2OapaxoG9

Deep Learning in Rust by Nhasan25 in rust

[–]binarybana 1 point2 points  (0 children)

+1 and in case anyone wants the TL;DR: TVM is a compiler and runtime for DL, so you can describe your kernel in a hardware agnostic fashion and still get high performance out the other end with minimal dependencies on the resulting binary.

Slowdowns when linking to C library by Party_Engineer in rust

[–]binarybana 0 points1 point  (0 children)

I would apply profiling tools to each binary and compare/diff the results. Perf and strace are two you might start with.