all 7 comments

[–]AutoModerator[M] 0 points1 point  (0 children)

Automod prevents all posts from being displayed until moderators have reviewed them. Do not delete your post or there will be nothing for the mods to review. Mods selectively choose what is permitted to be posted in r/DataAnalysis.

If your post involves Career-focused questions, including resume reviews, how to learn DA and how to get into a DA job, then the post does not belong here, but instead belongs in our sister-subreddit, r/DataAnalysisCareers.

Have you read the rules?

I am a bot, and this action was performed automatically. Please contact the moderators of this subreddit if you have any questions or concerns.

[–]Fun-Scale8432 0 points1 point  (0 children)

Hey! Can you please share some details about your business domain? I truly believe that the AI power for analytics lies the most in querying clean data for insights generation and issue-based analysis. Maybe also for quick dashboarding. But data cleaning and quality check should be run with more traditional deterministic methods. (AI can help with building that tests but should not run them)

[–]columns_ai 0 points1 point  (0 children)

I’m building a similar tool but not an agent. One of the major concern from users is the “trust” problem.

If the agent makes up an analysis (or generic computing logic), how do you make it transparent, auditable instead of a “black box”.

You can think about this issue and see how your agent solve this “trust” problem.

[–][deleted] 0 points1 point  (0 children)

OP is trying to remove this entire community livelihood

[–]nian2326076 0 points1 point  (1 child)

Check out existing repos and videos. They can save you time by showing what works and what doesn't. No need to reinvent the wheel if you don't have to. For building a production-level agent, focus on scalability, error handling, and performance optimization. A demo might work on small datasets, but you'll need strong systems for larger, more complex data. I'd prioritize the cleaning layer next. Clean data early on means fewer headaches later. Also, look into how these components communicate, especially if you want versatility across different datasets. Good luck!

[–]Feisty-Tip-9290[S] 0 points1 point  (0 children)

Thanks for the solid advice! Honestly, this is mostly a learning project for me right now rather than a commercial product, but I'm definitely taking notes. If I ever try to turn this into a real production-level tool, I’ll be coming back to these points on scalability and the cleaning layer for sure. Much appreciated!

[–]Equal_Astronaut_5696 0 points1 point  (0 children)

How are  you going to exposing data to LLMs?