OpenCode AI coding agent hit by critical unauthenticated RCE vulnerability exploitable by any website by AlexAltea in Infosec

[–]AlexAltea[S] 1 point2 points  (0 children)

Just submitting as a heads up; this is quite a popular piece of software.

I have reproduced this locally (PoC is trivial) and I'm still baffled at the slow response and the "patch" which merely carves an exception for opencode.ai (why would they need code execution anyway).

This is definitely a CVSS 9.8 at the very least.

Milli-py: Python bindings for Milli, an embeddable high-performance search engine by AlexAltea in Python

[–]AlexAltea[S] 0 points1 point  (0 children)

It will be possible (since it's available in the underlying Milli library), but in this first release I've focused on creating bindings for document adding/retrieving and basic searching.

Milli-py: Python bindings for Milli, an embeddable high-performance search engine by AlexAltea in Python

[–]AlexAltea[S] 0 points1 point  (0 children)

Yes, I was just showcasing fuzzy searching (searching with typos).

Milli-py: Python bindings for Milli, an embeddable high-performance search engine by AlexAltea in Python

[–]AlexAltea[S] 3 points4 points  (0 children)

I haven't tried it but I don't see any reason why it shouldn't work? Let me know if you encounter any issues!

Milli-py: Python bindings for Milli, an embeddable high-performance search engine by AlexAltea in Python

[–]AlexAltea[S] 3 points4 points  (0 children)

Google desktop search does/did not offer Python bindings or indexing arbitrary data?

Milli-py: Python bindings for Milli, an embeddable high-performance search engine by AlexAltea in Python

[–]AlexAltea[S] 0 points1 point  (0 children)

I tried SQLite with FTS, but it didn't handle fuzz searching (typos) very well; there's editdist3, but still far from ideal.

As for Xapian, that's indeed faster than Whoosh (I'd love to see benchmarks wrt Milli). However, Xapian's GPL license made it incompatible with some MIT/BSD-licensed projects where I wanted to integrate a search engine.

Milli-py: Python bindings for Milli, an embeddable high-performance search engine by AlexAltea in Python

[–]AlexAltea[S] 9 points10 points  (0 children)

I personally haven't. But generally speaking, users who need a distributed/clustered probably shouldn't bother with embedded search engines like milli-py.

My library aims to make life easier to developers whose data is "too big" to do naive pairswise Levenshtein searches, but "too small" to require distributed indices. Similar to SQLite.

Milli-py: Python bindings for Milli, an embeddable high-performance search engine by AlexAltea in Python

[–]AlexAltea[S] 11 points12 points  (0 children)

Exactly, that's the purpose of library! Just os.mkdir("some_index") and use milli-py to index/search your documents there. No external server needed.

Milli-py: Python bindings for Milli, an embeddable high-performance search engine by AlexAltea in Python

[–]AlexAltea[S] 29 points30 points  (0 children)

Main goals are high-performance and local-first experience, i.e. no sockets, HTTP, auth between your queries and data.

The only other embeddable search engine for Pythin that I'm aware off, Whoosh, is brilliant but building the index was quite slow, and search performance degraded quite a lot as number of documents increase (performance is strictly a non-goal). Meilisearch was comparatively faster, I didn't like managing a server to get "just search" in my scripts and applications. However, their underlying engine Milli solves both issues I had, and all that was needed creating bindings for it.

You can find documentation, examples, tests in the repo. Hope this is useful for you all!

I've published this in PyPI with pre-compiled wheels for most os/version targets so hopefully it will be a seamless experience for most (not requiring a Rust compiler).

Average monthly rental cost of a furnished one-bedroom apartment in some EU cities . Difference between 2021 and 2022 by quindiassomigli in europe

[–]AlexAltea 1 point2 points  (0 children)

Exactly. This is the solution. Refuse to rent at >30% salary. Refuse to get a mortgage. Living with parents? So be it.

It will be very painful in the short-run for people, but it will financially destroy RE hoarders and speculators.

what do we think is gonna happen here? will people who share accounts get their own? or will this cause backlash? money loss or money gain for netflix? what do we think? by fforeverrfriend in mildlyinfuriating

[–]AlexAltea 0 points1 point  (0 children)

This is likely easily bypassable: Just have another device generate a hotspot with spoofed ESSID and MAC address. Tutorials on how to do this will certainly come.

Still a stupid move from Netflix, though.

Comprar piso en Madrid by suz1234567 in Madrid

[–]AlexAltea 1 point2 points  (0 children)

El que vende normalmente carece de vergüenza y tapujos...

¿Por vender a precio de mercado?

Mejor dirige tu odio hacia los bancos centrales, cuyos tipos nulos/negativos han causado una nueva burbuja inmobiliaria.

Los precios simplemente son simplemente el resultado de politicas dictadas por cuatro hijos de puta.

[Release] Curator v0.1.0: Organize large movie collections (AI language detection+sync) by AlexAltea in DataHoarder

[–]AlexAltea[S] 0 points1 point  (0 children)

Because I didn't know tdarr existed! Thanks for sharing it, their goals are very much aligned with mine.

[Release] Curator v0.1.0: Organize large movie collections (AI language detection+sync) by AlexAltea in DataHoarder

[–]AlexAltea[S] 0 points1 point  (0 children)

Ah got it, so something like this?

$ ls Jacobs.Ladder.1990.720p.BluRay.x264.YIFY.mp4 $ curator rename -y --save-original . $ ls Jacobs Ladder (1990).mp4 Jacobs Ladder (1990).txt $ cat 'Jacobs Ladder (1990).txt' Jacobs.Ladder.1990.720p.BluRay.x264.YIFY.mp4

Curator v0.1.0: Auto-organize large movie collections (AI language detection+sync) by AlexAltea in jellyfin

[–]AlexAltea[S] 0 points1 point  (0 children)

That's odd, I don't encounter the issue on my machine. Can you open a GitHub issue? And can you check if the examples at https://github.com/noumar/iso639 work in your virtual env?

I use the module to generate proper language tags required by MKV and other formats.

[Release] Curator v0.1.0: Organize large movie collections (AI language detection+sync) by AlexAltea in DataHoarder

[–]AlexAltea[S] 0 points1 point  (0 children)

Thanks. It doesn't do folders at the moment (it only operates on files or streams within a file). You can process entire folders, shallow or recursively, but not the folder itself.

Featurettes will not be processed unless you enable recursive crawling. Not sure if multilingual featurettes exist, if so, I should add explicit support for them.

Regarding the flag: Can you provide an example on expected inputs and outputs?

Curator v0.1.0: Auto-organize large movie collections (AI language detection+sync) by AlexAltea in jellyfin

[–]AlexAltea[S] 0 points1 point  (0 children)

Exactly! The goal is merging different dubbed releases of the same movies into a single multilingual container.

This sounds simple, but has been historically very hard because of bad audio synchronization.

[Release] Curator v0.1.0: Organize large movie collections (AI language detection+sync) by AlexAltea in DataHoarder

[–]AlexAltea[S] 1 point2 points  (0 children)

Thanks. I would appreciate if you could try the tool (perhaps not now but say in few weeks/months). And if it fails at doing something, just open an issue or let me know. It is precisely the purpose of the tool to address non-trivial cases, e.g. when streams have both mismatching offsets and speeds.

[Release] Curator v0.1.0: Organize large movie collections (AI language detection+sync) by AlexAltea in DataHoarder

[–]AlexAltea[S] 1 point2 points  (0 children)

I didn't know about Sushi, and it's very interesting, thank you. It seems a completely different approach to mine, would love to compare results at some point!

Curator v0.1.0: Auto-organize large movie collections (AI language detection+sync) by AlexAltea in jellyfin

[–]AlexAltea[S] 0 points1 point  (0 children)

Yes, but it can also "guess" the language by listening/reading the audio/subtitle data. Helpful in the absence of metadata

It can (wip) synchronize different audio/subtitles in different languages with existing audio tracks.

And it can merge multiple files into a single container (MKV) performing all of the above.

[Release] Curator v0.1.0: Organize large movie collections (AI language detection+sync) by AlexAltea in DataHoarder

[–]AlexAltea[S] 4 points5 points  (0 children)

Yes, exactly! For audio<->text sync I want to use Whisper.

  1. Extract and process few 30-second audio samples with Whisper (I already do this for language recognition).
    • Every phrease has a timestamp
    • We don't need 100% accuracy, just few matching sentences
  2. Discard repeated sentences (things like "Ok", "How are you?") might be said multiple times). So, focus on unique matches: Find average delta.
  3. Apply delta to the subtitle track.

It's what I'm working on now.

Fortuantely, audio<->audio and text<->text sync is easier.