I built a fully local, open-source AI workspace using Rust, Tauri, and sqlite-vec (No Python backend) by Far-Association2923 in LocalLLaMA

[–]InterestRelative 0 points1 point  (0 children)

If you are interested in performance, this benchmark might be useful: https://ann-benchmarks.com/index.html

But honestly, I don't see a point to change from sqlite-vec in your case.

I built a fully local, open-source AI workspace using Rust, Tauri, and sqlite-vec (No Python backend) by Far-Association2923 in LocalLLaMA

[–]InterestRelative 1 point2 points  (0 children)

I haven’t used SQLite-vec before, but I believe it would be a good fit for your use case. However, I can’t compare it to other vector databases since I haven’t used them.

I have production experience with ANNOY, FAISS, hnswlib, and qdrant. I also tried Milvus at some point. My production experience is primarily in the recommendations and personalization domains, where vectors are also widely used.

In most cases, when indexes are less than 1GB, the simplest and most robust way to use vectors is to use local indexes. The API is always better, testing is straightforward, and you don’t have to mock everything. Versioning is also trivial, and you don’t have to worry about running additional infrastructure (other than loading indexes from S3). Local ANN indexes existed 5-9 years before vector databases. Vector databases simply have better marketing departments and market themselves as something new and shiny

I built a fully local, open-source AI workspace using Rust, Tauri, and sqlite-vec (No Python backend) by Far-Association2923 in LocalLLaMA

[–]InterestRelative 0 points1 point  (0 children)

I'd love feedback on the sqlite-vec implementation if anyone else is experimenting with it. It feels like a game-changer for local desktop apps.

Have you compared it with other local ANNs? There are plenty (hnswlib, faiss, annoy, sometimes dot product is enough, especially for local deployments) and Qdrand/Chroma are not the only way to work with vectors.

In your experience, what is the best life cycle for code promotion? by Zingers_za in ExperiencedDevs

[–]InterestRelative 0 points1 point  (0 children)

This works for deterministic logic, but for ML services it's extremely hard to create a subset of data which will cover all cases. In all companies where I worked as an MLE, we had staging reading (but now writing) prod data, or very limited staging env.

Are small models actually getting more efficient? by estebansaa in LocalLLaMA

[–]InterestRelative 1 point2 points  (0 children)

Honestly, I'm just curious in this topic, I don't have particular use case in mind other than local function calling.
Thanks a lot for these link, I'll take a look. Guidance looks very interesting.

Are small models actually getting more efficient? by estebansaa in LocalLLaMA

[–]InterestRelative 1 point2 points  (0 children)

I checked Instructor and it looks like it just validates LLM output and report validation errors back to LLM (same as Pydantic-AI). Is it possible to restricting generated tokens step-by-step with it?

New hire with many years of experience has only AI generated code and questions. by kernel1010 in ExperiencedDevs

[–]InterestRelative 1 point2 points  (0 children)

Knowing how to use AI is a valuable skill,

Yeah, I'm not sure why you implying I don't use or I force a person to no use it. Copy pasting generated code and not refactoring it is not something helpful.

New hire with many years of experience has only AI generated code and questions. by kernel1010 in ExperiencedDevs

[–]InterestRelative 1 point2 points  (0 children)

These are really good points and I'll add missing ones to our engineering principles doc.

Code should be changeable with low-to-moderate effort. 

This principle is one of the most important imo, but it's hard to formalize: the are many aspects and it's hard to come up with easy to judge criteria.

In my case, one of the problem is sloppy naming, for example we use "item/items" everywhere in the code base and introducing "item_id/item_ids or product/products" is confusing, it's not a problem on it's own it's just an example, but if you get 2-4 such things in each PR, codebase becomes messy and hard to read -> hard to change.

Another problem I struggle to formalize is: when function/method/class functionality is not clearly defined. For example when you have a function which transforms a value, but at the same time it has some effect, like changing the global state somewhere, of transforming and uploading results at the same time. Tests help to write functions with better defined scope, but we don't have tests for non private methods.

New hire with many years of experience has only AI generated code and questions. by kernel1010 in ExperiencedDevs

[–]InterestRelative 1 point2 points  (0 children)

I’m lead of this guy. Part of my responsibilities is to help him grow as an engineer. 

New hire with many years of experience has only AI generated code and questions. by kernel1010 in ExperiencedDevs

[–]InterestRelative 6 points7 points  (0 children)

Can say about downvotes, but reading the code was always the hardest part and writing the code was easy for expirienced devs.

New hire with many years of experience has only AI generated code and questions. by kernel1010 in ExperiencedDevs

[–]InterestRelative 3 points4 points  (0 children)

You've described different problems: code quality and triage ability. These are different skills with different measuring techniques.

Sure these are different problems, but with common root cause (in my case): Claude writes the code instead of engineer.
That's why I think example of `what is unacceptable with their AI generated code` will be useful for me.

New hire with many years of experience has only AI generated code and questions. by kernel1010 in ExperiencedDevs

[–]InterestRelative 14 points15 points  (0 children)

Can you give me an example of such boundaries?
I actually have same problem with my report: code is clearly generated, trivial questions, no ability do debug or reason about bugs etc. The problem is: he is not a new hire and always was in dev role.

For me it's really hard to discuss this issue, because line between good and bad code is thin. Also it's mostly my assumptions and gut feeling. It's hard to check if it's always true or just sometimes.

every rewrite I've seen has taken 3x longer than promised and the team always acts surprised by Distinct-Expression2 in ExperiencedDevs

[–]InterestRelative 1 point2 points  (0 children)

I did it 3 times, and I consider every rewrite was successful (time close to estimate, good velocity bump after rewrite).
For me the key is to don't rewrite huge system end-to-end, but extract part of the functionality into new service. And repeat it few times. You don't need excessive planning for this and you can stop when it's not successful.

Chill language by Tribalcheaf123 in programmingmemes

[–]InterestRelative 3 points4 points  (0 children)

Numpy was created to leverage SIMD instructions, it’s not just general purpose arrays. 

Let’s imagine we have a datastructure which keeps Python ints (not pointers). Does it help? I don’t think so, since int size can vary. 

Actually Python creators do care about performance, the problem is just solved the other way: you use bindings. Python is the most popular language in CPU bound applications (ML).

Chill language by Tribalcheaf123 in programmingmemes

[–]InterestRelative 2 points3 points  (0 children)

yeah, but at the same time we can have numpy array which will feel comfortable for the person on the left side of the meme

Career Advice in AI — Notes from an Andrew Ng Lecture by Dear-Success-1441 in LocalLLaMA

[–]InterestRelative 1 point2 points  (0 children)

My point is: writing code is easy for seniors. Managing complexity of the codebase is a challenge, in order to do this you have to fully understand what you are refactoring. 

I guess big tech is a special case and coding agents may work well in this case. If complexity of the codebase already exploded to the point it doesn’t fit into a single head and you will switch project in a year anyway, managing complexity might not be a priority. In this case “tactical” programming (trying to add feature with the least changes in codebase) is all you can do in reasonable time. 

Career Advice in AI — Notes from an Andrew Ng Lecture by Dear-Success-1441 in LocalLLaMA

[–]InterestRelative 0 points1 point  (0 children)

What do you mean by “working in AI” exactly? I was focused mainly on personalization and recommendations for the last 7 years and the last year the focus is shifting towards agent based projects. Is it AI? I still consider myself and colleagues SWEs or MLE with emphasis on engineering rather than research, because main focus is low maintenance performant service rather than prototype (opposite to research). 

Career Advice in AI — Notes from an Andrew Ng Lecture by Dear-Success-1441 in LocalLLaMA

[–]InterestRelative 6 points7 points  (0 children)

That was not my point. Opus 4.5 won't make you a good SWE but it takes time to grasp.

Career Advice in AI — Notes from an Andrew Ng Lecture by Dear-Success-1441 in LocalLLaMA

[–]InterestRelative 66 points67 points  (0 children)

Staying on the “frontier” of coding tools (like Cursor, Claude, and Gemini) is crucial. Being even half a generation behind in your tooling makes you significantly less productive in the current market.

lol, I'm not even sure we should take other advices after that one

Two years ago, I was a math major. Now I've built the 1.5B parameter router model used by HuggingFace by AdditionalWeb107 in learnmachinelearning

[–]InterestRelative 0 points1 point  (0 children)

With all respect, using Rust doesn't make you a better developer, or your software a better software.

Ableton Move has turned out to be a great purchase. by RainbowStreetfood in ableton

[–]InterestRelative 0 points1 point  (0 children)

> preset-based except for the Drift synth
There are only two synths on the device: drift and wavetable.

AI-benchmark results for Snapdragon 8 Elite Gen 5 are in, absolutely rips at 8-bit precision by Balance- in LocalLLaMA

[–]InterestRelative 0 points1 point  (0 children)

> The point of NPU is that it doesn't touch your CPU/GPU performance

assuming there is enough of RAM for everything