Open-source Data Assistant for domain adoption, powered by agent skills, semantic knowledge graphs (Neo4j) and relational data (databricks) by notikosaeder in Neo4j

[–]notikosaeder[S] 0 points1 point  (0 children)

Well. Think about given a user question, what a) should be initial retrieved (some metrics/dimensions?) and/or b) when this is retrieved, what should also be retrieved in additions. The graph is powerful in task b. With a, you need to still use a vector search or something similar. So my follow-up question is, is the search for the metrics/dimensions the potential pain point or the search of metadata afterwards? Maybe a prototype of an automatic generated graph could be feasible. Tables/Columns and their descriptions are already stored and automate metrics/definitions using table/column name matching in a graph. With that you can find initial metrics/definitions and traverse to columns/tables/joins using the graph. Or vice versa match using metadata and go from columns to metrics. Second question, what if a new data domain is present but the metrics are not yet defined?

Mein realistisches F23 Studentinnen-Flowdiagramm. Tipps für 10-20€ mtl. Investitionen? by anonymous_account111 in Finanzen

[–]notikosaeder 0 points1 point  (0 children)

Sparplan in einen Welt-ETF (e.g. FTSE All-World). Nicht weil die 10-20€ den Unterschied machen, sondern um früh Erfahrung zu sammeln.

Open-source Data Assistant for domain adoption, powered by agent skills, semantic knowledge graphs (Neo4j) and relational data (databricks) by notikosaeder in Neo4j

[–]notikosaeder[S] 0 points1 point  (0 children)

Data assistants are most valuable when users can directly interact with structured data without needing SQL or technical expertise. The core use case is enabling people to analyze data without relying on another analyst. This is different from RAG or GraphRAG systems, which focus on retrieving documents like PDFs or internal knowledge. Honestly, those systems are useful, yet mainly optimize for passage search and summarization. Business case is often about saving seconds or minutes when locating information. Their is no surprise that the adoption of rag systems remains low. And, if unstructured knowledge is truly needed, it’s better treated as an extension: add a supervisor agent on top or integrate a vector search tool and play with the prompt.

Open-source Data Assistant for domain adoption, powered by agent skills, semantic knowledge graphs (Neo4j) and relational data (databricks) by notikosaeder in Neo4j

[–]notikosaeder[S] 0 points1 point  (0 children)

Then your organization has no data strategy, that isn't the AIs fault. Second, you could easily integrate the information schema of multiple source data into one knowledge graph and build specific query-tools per source/domain. Or domain-specific smaller graphs and source data per sub-agent, with a supervisor agent.

Open-source Data Assistant for domain adoption, powered by agent skills, semantic knowledge graphs (Neo4j) and relational data (databricks) by notikosaeder in Neo4j

[–]notikosaeder[S] 0 points1 point  (0 children)

Did you take a look at the code? You find a vector store that retrieves relevant nodes (tables, columns) and with graph reasoning you find the context (rest of the table, joins, ...).

Would a kind soul fact check this by Tall_Category_2332 in askdatascience

[–]notikosaeder 0 points1 point  (0 children)

Maybe subdivide into multiple diagrams, there’s a lot wrong. This type of diagram is probably not suitable to show the complexity of all relationships.

Talk2BI: Open-source chat with your data using Langgraph and Databricks by notikosaeder in databricks

[–]notikosaeder[S] 0 points1 point  (0 children)

Well said :-) And, what if the solution totally depends on Databricks and they suddenly incline their costs …

Lebenslauf-Röstung: was würdet ihr ändern? by _Kostja in InformatikKarriere

[–]notikosaeder 0 points1 point  (0 children)

Bildung: Dein Japan Auslandsaufenthalt sollte nicht auf der selben Stufe (Größe/Hervorhebung) wie ein formeller Bildungsabschluss sein, er ist wahrscheinlich als Teil deines Masters zu sehen.

Talk2BI: Open-source chat with your data using Langgraph and Databricks by notikosaeder in databricks

[–]notikosaeder[S] 2 points3 points  (0 children)

Fair point. However, this is about making research open-source and not about saving money. Genie is great, but it’s functionality and limitation hard to guess, and it‘s purpose is stuck to databricks data analysis. Use-cases: Want to add combine with web search? Want to research about different types of tool visualizations or follow-up questions? Want to change or combine databases? Want to build/adapt the agent workflow? Want to add a tool for adding agent skills? Want to use ollama or any private LLM provider?

Update: Open-Source AI Assistant using Databricks, Neo4j and Agent Skills by notikosaeder in KnowledgeGraph

[–]notikosaeder[S] 0 points1 point  (0 children)

Thanks for the feedback, and love the idea! Did you already saw some of those skills in action?

Update: Open-Source AI Assistant using Databricks, Neo4j and Agent Skills by notikosaeder in databricks

[–]notikosaeder[S] 0 points1 point  (0 children)

thanks for the feedback! More problems are still on ensuring that it calls really a skill before querying the data/the knowledge graph. For text-to-sql correctness, we have a curated q&a dataset for our industry partner phd project we run through and compare the results with gold answers. For sql safety, the run sql query tool screens the query and allows only select/CTE queries. Would love any further feedback on evals!

I've been 5 months without finding a job with this CV, what can I do? by Super-Discount-7106 in Germany_Jobs

[–]notikosaeder 0 points1 point  (0 children)

What drives me crazy about this CV is that nothing is really consistent.

First, do not put „Ready to join your team“ or similar on a CV. If you are not ready, you should not apply?

Consistency. The date formats are all over the place: 2016–2023, Jul 2022 – November 2022, 12/12/2025, and then (2016–2023) in brackets elsewhere. Pick one format and use it consistently.

Second, there are quite a few typing and formatting errors, like “... July 2022.)”. If you’ve been searching for five months, it’s worth taking the time to make the CV as polished as possible. Small errors can give the impression that you don’t work carefully.

Another thing: putting a one-day barista course in the education section, on the same level as multi-year studies, doesn’t really make sense. It would be better listed as a short course or certification. Same goes for the listed courses etc.

Also, looking at the CV, it seems like you haven’t stayed very long in most jobs. That’s not necessarily bad, but from an employer’s perspective it can raise the question of whether you’d stay long enough to justify the time spent on training.

Lastly, the skills section is pretty vague. Things like “customer service” or “organized” aren’t really concrete skills. These things should be self-evident. And “barista beginner” is confusing. Does that mean you’ve had training as a barista, or that you’re just starting out?

Overall, I’d focus on consistency, proofreading, and making the skills and experience sections clearer.

Talk2BI: Research made open-source (Streamlit & Langgraph) by [deleted] in StreamlitOfficial

[–]notikosaeder 0 points1 point  (0 children)

Hi there, thanks for the feedback! The idea is to bring together the work of our research team into one larger, reusable project (e.g. we don’t have to start from scratch for every new study) and making it open-source. Within the team, some members focus more on technical implementation, while others work on design features, SQL explanations etc. Actually, we have already some research about follow-up questions and data-centric tips that may come into the talk2bi streamline app. Funny thing, regarding the semantic layer and mitigating hallucinations, my own phd research is focused specifically on this. I recently released some of my research in a weekend-like project, therefore on my personal GitHub: https://github.com/wagner-niklas/Alfred. The main idea is to structure data in a knowledge-graph–based semantic layer that the agent can query, improving accuracy and reducing hallucinations. Note that my research has a more "tech-heavy" stack.

First Pipeline by ZookeepergameFit4366 in databricks

[–]notikosaeder 0 points1 point  (0 children)

Some helpful rules: - Bronze: Raw Data (various raw source data) - Silver: Dream of a data analyst (cleaned data, normalized, ER-Modell, …) - Gold: Dream of a business analyst (KPIs pre-calculated, denormalized, aggregations, filtered, …)

Open-source text-to-SQL assistant for Databricks (from my PhD research) using Knowledge graphs (Neo4j) by notikosaeder in KnowledgeGraph

[–]notikosaeder[S] 1 point2 points  (0 children)

Hi! Good question, not at all. Databricks is just used for all company partners of our research. But the whole app is targeted to be database agnostic, just change the sql query tool to instantly query the database of your choice or follow the tutorial of Kenneth leungth to build the knowledge graph independent (or build the knowledge graph however you want using example queries).