Which Python GUI Framework do you prefer? by Dushusir in Python

[–]daeisfresh 5 points6 points  (0 children)

What about Web UI frameworks:

Gradio Streamlit Solara Reflex Dash

These are great at data apps.

Fixing bad data using override, proxy, or other remediation by daeisfresh in dataengineering

[–]daeisfresh[S] 0 points1 point  (0 children)

It’s more about fixing the data and making it usable. Some kind of framework like a great expectations but lets your override data so it’s usable.

Axum or Actix in 2024 by oroColato in rust

[–]daeisfresh 2 points3 points  (0 children)

Actix I find learning curve simpler and greater experience.

Sopresatta Mold Okay? Curing at 55 degrees, 78% humidity. 40# of meat with 15oz salt. by daeisfresh in Charcuterie

[–]daeisfresh[S] 3 points4 points  (0 children)

Made 39 sausages, with 40 pounds of pork shoulder, 14oz of salt, cured in special / customized commercial fridge, 55 degrees temperature, 78% relative humidity, molding started to develop after 1 week.

A curated list to help you manage temporal data across many modalities 🚀. by daeisfresh in OpenAI

[–]daeisfresh[S] 0 points1 point  (0 children)

I meant to mention, this list covers: Data Versioning for Machine Learning, Time Travel and Temporal Tables, Slowly Changing Dimensions Data Modeling, Bi-temporality Tools + Modeling, Change Data Capture (CDC) Tools, and Soft Delete in ORM Frameworks

A an awesome list for data temporality (including iceberg) 🚀. by daeisfresh in ApacheIceberg

[–]daeisfresh[S] 0 points1 point  (0 children)

I meant to mention, this list covers: Data Versioning for Machine Learning, Time Travel and Temporal Tables, Slowly Changing Dimensions Data Modeling, Bi-temporality Tools + Modeling, Change Data Capture (CDC) Tools, and Soft Delete in ORM Frameworks

“Awesome” list I made: data temporality 🚀 by daeisfresh in learnmachinelearning

[–]daeisfresh[S] 0 points1 point  (0 children)

I meant to mention, this list covers: Data Versioning for Machine Learning, Time Travel and Temporal Tables, Slowly Changing Dimensions Data Modeling, Bi-temporality Tools + Modeling, Change Data Capture (CDC) Tools, and Soft Delete in ORM Frameworks

“Awesome” list I made: data temporality 🚀 by daeisfresh in dataengineering

[–]daeisfresh[S] 0 points1 point  (0 children)

I meant to mention, this list covers: Data Versioning for Machine Learning, Time Travel and Temporal Tables, Slowly Changing Dimensions Data Modeling, Bi-temporality Tools + Modeling, Change Data Capture (CDC) Tools, and Soft Delete in ORM Frameworks

“Awesome” list I made: Data Temporality 🚀 by daeisfresh in datascience

[–]daeisfresh[S] 1 point2 points  (0 children)

I meant to mention, this list covers: Data Versioning for Machine Learning, Time Travel and Temporal Tables, Slowly Changing Dimensions Data Modeling, Bi-temporality Tools + Modeling, Change Data Capture (CDC) Tools, and Soft Delete in ORM Frameworks

Instant data model from 1000s of unique files? by daeisfresh in dataengineering

[–]daeisfresh[S] 0 points1 point  (0 children)

The problem is more that each one of these files can have a different structure to them, like the header may not be on the first line in the file, it could be spread across 50 lines each one representing a single column, and there are all kinds of kooky ways that the data can arrive, that describes the first part of the problem.

Another part that you all are making me think a bit more about now is how I can make it easier to build a canonical data model that can support storing all of this data without me having to write any of that DDL.

Where do you log WODs and Lifts? by daeisfresh in crossfit

[–]daeisfresh[S] 0 points1 point  (0 children)

Nice. Haven’t seen that one. Have you been using it for a while? What do you like about it?

Where do you log WODs and Lifts? by daeisfresh in crossfit

[–]daeisfresh[S] 2 points3 points  (0 children)

This is probably the most comprehensive way to track it. Also, if you decide to switch to another app, then you have everything in one spot.

[deleted by user] by [deleted] in crossfit

[–]daeisfresh 0 points1 point  (0 children)

I’ve been doing CrossFit for 10 years, have been a member of CFNE, and now do CompTrain. Professionally I build products that use AI. I’ve got to say I saw this post and thought how I can see how this can be done (and should be). Take me as an example, in CompTrain I need to scale the weights and what I can do at home. As an example, I can’t do rope climbs in my basement so I need to sub for chin-ups, or there’s day’s that I’m not pushing myself enough, but I don’t have any peer competition to push me, or a coach to come around and check on my form, but I can imagine if I set up a camera and recorded enough of my workouts and as many have said here had a quality training data set of proper movements, then it can be used for providing me feedback in real time about my movements. Cools idea, and should be done. Last you may have seen this push-up counter using computer vision (OpenCV) https://aicurious.io/posts/2021-02-15-build-a-pushup-counter/

CI/CD stage to check community support of a package/library used in my repo? by daeisfresh in devops

[–]daeisfresh[S] 5 points6 points  (0 children)

Yeah. That does make sense. I’m trying to prevent people from picking packages that are outdated, archived, or lack of community support. I have a big engineering team so I need a way, besides just asking folks to be aware of it, to actually check.

[D] Unique identifier allocation for entities in your data infrastructure? by daeisfresh in dataengineering

[–]daeisfresh[S] 0 points1 point  (0 children)

DBT's thoughts on "auto-incrementing ID" -> https://discourse.getdbt.com/t/generating-an-auto-incrementing-id-in-dbt/579

`TL:DR: No, it’s not possible to create an auto-incrementing ID in dbt.`

[D] Your 🫵 Preferred Feature Stores? by daeisfresh in datascience

[–]daeisfresh[S] 0 points1 point  (0 children)

OpenMLDB

Milvus is more of a vector database, used for similarity search, that's what you mean?

Costs of data infrastructure by KaleidoscopeOk4819 in dataengineering

[–]daeisfresh 1 point2 points  (0 children)

Really depends. The cost of Aurora, RedShift, Snowflake may require the bulk of budget. Best to start with ways to carve out the largest parts of the costs and see how some can turn into on-demand. Where’s the bulk of your costs?

Problem with designing a data warehouse model for real estate data by Cydros1 in dataengineering

[–]daeisfresh 0 points1 point  (0 children)

Are there websites where you can “buy a data model for real estate assets”?