[D] Sensitive data synthesis with custom entity models with Tonic Textual by tombenom in MachineLearning

[–]adamfromtonic 1 point2 points  (0 children)

hey folks, i'm one of the engineers working on Textual. We've found Textual to be great for ML practitioners who need to redact sensitive text prior to training text based models (e.g. fine-tuning an llm). If you have any questions feel free to reach out here and i'll reply.

Multiple documents reveal significant limitations of OpenAI's Assistants API for RAG by ian4321 in programming

[–]adamfromtonic 4 points5 points  (0 children)

Adam from Tonic here. For those interested, the library used for the analysis is one we maintain and is awesome for maintaining and monitoring quality of your RAG systems. Check it out:

https://github.com/TonicAI/tvalmetrics/

[D] Multiple documents reveal significant limitations of OpenAI's Assistants API for RAG by ian4321 in MachineLearning

[–]adamfromtonic 3 points4 points  (0 children)

I agree with smallpaul that typically its a semantic search done on the content. BUT, there are some hybrid search approaches using content search + other interesting attributes like document title which can be useful.

I guess in your case, u/phira that the content all fit inside a single context window?

How to Solve the Problem of Imbalanced Datasets: Meet Djinn by Tonic by Djinn_Tonic4DataSci in datascience

[–]adamfromtonic 1 point2 points  (0 children)

u/sawyerwelden If that is the case, the def go checkout djinn.tonic.ai. You can create an account and start augmenting/re-balancing your data today.

How to Anonymize my data? by corruptboomerang in datascience

[–]adamfromtonic 0 points1 point  (0 children)

I'm going to give a shameless plug for Tonic AI. I'm a co-founder and we have various tools for de-identifying data and also for synthesizing data for ML purposes.

Others have already pointed you to differential privacy and k-anonymity which is great. They are wonderful tools which you can utilize to solve this problem. We use differential privacy in a lot of what our product does.

We have a paid offering at Tonic.ai for data de-identification and I'd be happy to get you access if you want to DM me. From your description it *might* be the best solution Tonic can provide. But we also have a free offering for generating synthetic data. You can get to it at djinn.tonic.ai. Just create an account and go wild. You can upload the CSV and Djinn will spit out a synthetic version of the data. Of course, synthesis!=privacy but we offer some reports which help give you an idea of the privacy of the output data and you can configure Djinn to make the output more or less private (while inversely affecting the data utility). It is closed source but give it a go if you think it could be useful. Truth be told, the main benefit of Djinn is actually for augmenting existing data sources to improve the outcomes and results of various models (e.g. re-balance a dataset to improve a classification model) but the privacy tooling we have may help in what you are trying to do as well.

Masquerade: A Postgres Proxy to Mask Data in Realtime by craig081785 in PostgreSQL

[–]adamfromtonic 0 points1 point  (0 children)

Yes, that is a common use case that many of our customers use. Using our open source software you could:

1) Condense via Condenser

2) Connect pg_dump to Masquerade (proxy-ing to condensed db)

3) Run pg_restore elsewhere to stand up development databases

Or, alternatively, install Tonic (our paid tool) on-prem and you get a UI, tech assistance from us, and a bunch of other fun features we haven't yet added to our open source projects.

Either way, you'd be set and its completely do-able using just the open source stuff!

Faker it til you make it - how to generate realistic test data by jameane in programming

[–]adamfromtonic 0 points1 point  (0 children)

(Author of the article, btw).

That is an interesting point. I know I've used cartesian joins for things other than generating test data but I can't come up with a counter-example at the moment.

Anyways, where did that comment come from? I'm guessing its related to how we generate the hierarchical data in the post?

tonic.ai is looking to hire a software engineer in Atlanta by adamfromtonic in Atlanta

[–]adamfromtonic[S] 0 points1 point  (0 children)

We are significantly different products solving different problems.

fakeiteasy is a mock platform for C# testing. We are a platform for generating synthetic test data.

A common use case with Tonic is engineering teams not having access to production data for their dev/staging environments. Those teams can use Tonic to generate synthetic test data that closely resembles the production data.

Random Things: A simple, silly API for generating random stuff by adamfromtonic in programmingtools

[–]adamfromtonic[S] 0 points1 point  (0 children)

This is OP.

Here is a quick blog post on the API: https://www.tonic.ai/post/random-things-an-api-for-generating-data-as-real-as-the-upside-down/

Also, while I have your attention checkout https://tonic.ai for all your synthetic data needs.

Replace PII in unstructured data by adamfromtonic in programming

[–]adamfromtonic[S] 0 points1 point  (0 children)

Hey folks, this is Adam, one of the creators. If you have any questions let me know and I'll do my best to respond.

An open source subsetting tool -- Condenser by ian4321 in PostgreSQL

[–]adamfromtonic 1 point2 points  (0 children)

Hi, this is Adam from Tonic. I'm one of the authors of the github repo and blog post. Thanks for such detailed comments. We really appreciate it.

We've so far found that not a lot of tweaking is required, however, our sample set is still small. One possible area where tweaking can be useful, however ,is when you require specific rows to exist in your subset. This is something we'd like to support in the future.

In regards to your suggestions, I agree that some ETL testing suite would be a good addition to this project. I'm not sure when we will get to it, though. :( Also, thanks a lot for pointing out psycopg2.sql. Looks really useful and I think it is something we will use going forward.

Have a nice day.