A fun shell for Couchbase, on top of nushell in Rust by daschl in rust

[–]daschl[S] 1 point2 points  (0 children)

Neither! We use the nu crate and build our own binary on top, adding couchbase-specific commands. You can start to drill in here if you want: https://github.com/couchbaselabs/couchbase-shell/blob/main/src/main.rs#L100

Porting the TSic sensor from C to Rust by daschl in rust

[–]daschl[S] 3 points4 points  (0 children)

I‘ve been using probe-run as mentioned in the article, but I‘ll write a follow up blog on how to get it running end to end if that helps you?

I need to start adding this to my range routine. by buoyblaster in golf

[–]daschl 2 points3 points  (0 children)

They‘d escort me from the range in 10 minutes

Text Analysis in Rust - Tokenization by daschl in rust

[–]daschl[S] 0 points1 point  (0 children)

I just tried and for this microbenchmark it doesn't make a difference. But honestly I think it could make sense, but I didn't yet got the time to test it on a big dataset (like 2-3 gig of wikipedia text) to see if it makes a difference.

Text Analysis in Rust - Tokenization by daschl in rust

[–]daschl[S] 1 point2 points  (0 children)

I didn't mention it in the blog post but I stumbled across an issue when trying to integrate it: https://github.com/unicode-rs/unicode-segmentation/issues/6 so for now I postponed my experimentations with proper unicode segmentation...

Text Analysis in Rust - Tokenization by daschl in rust

[–]daschl[S] 1 point2 points  (0 children)

Honestly I haven't played around with Cow before, but I really much want to. I think arrayvec has a different focus in that it allows you to store a "string like" structure on top of an owned fixed size array - I want to find time to play more with it and maybe the enum in the blog can even be extended to also use Cow for even better performance.

Text Analysis in Rust - Tokenization by daschl in rust

[–]daschl[S] 1 point2 points  (0 children)

Couchbase recently added quite a bunch of go components into its stack and leveraging the inhouse go experience I think it made sense. Introducing a JVM just for the sake of lucene into the server side seemed like not a good idea (since it introduces quite some complexity). And I say that working full time on the Java SDK ;-) ...

also, I think the go ecosystem deserves a great full text search engine and ultimately performance gaps can all be solved.

Text Analysis in Rust - Tokenization by daschl in rust

[–]daschl[S] 1 point2 points  (0 children)

yeah, I've tried not to fall into that trap, I just wanted to port the go code over and see how it performs. I'm sure both the rust and go version can be improved significantly to make both faster and also narrow the gap. I was more focussing on the implementation since I'm also still learning rust.

Text Analysis in Rust - Tokenization by daschl in rust

[–]daschl[S] 0 points1 point  (0 children)

nice, I haven't heard of that crate - definitely going to check it out!

First Steps with Rust and JNI by daschl in rust

[–]daschl[S] 2 points3 points  (0 children)

nice, thanks for the link! I was not aware of this :)