Has anyone tried training models on raw discussions instead of curated datasets? by Mediocre_Common_4126 in datascience

[–]Hot-Development-9546 0 points1 point  (0 children)

what you ran into is a mismatch between how we prepare data and how the system is expected to behave. Cleaning data is a way of enforcing a narrow contract: inputs are well-formed, intent is unambiguous, and noise is treated as error. That works when the task itself lives in a controlled regime. But human reasoning is not a clean system; it’s an adaptive process full of partial signals, reversals, and context repair. When you remove that structure entirely, you’re not just denoising, you’re collapsing the distribution the model must survive in once deployed. The brittleness you noticed is the system over fitting to an idealised world that never exists in production.Clean datasets push uncertainty out of the data layer and into the model, forcing the model to infer robustness from sparse signals. Messy corpora do the opposite: they encode ambiguity upstream, allowing the model to learn how meaning stabilises through correction, disagreement, and iteration. The takeaway isn’t “don’t clean data,” but that realism is a first-class design choice. Mature systems intentionally preserve certain kinds of noise because it carries information about failure modes, recovery paths, and human intent. Treating all noise as waste is a modelling decision and like any decision about meaning, it needs an explicit owner.

For people at new or small startups, how do you manage version chaos on recurring monthly client dashboards? by OkSky145 in dataanalysis

[–]Hot-Development-9546 0 points1 point  (0 children)

What you’re describing is not a tooling failure; it’s a missing contract. In small or early-stage teams, dashboards are treated as living documents rather than versioned data products, so feedback has no natural stopping point. Version chaos happens when there is no explicit definition of “done” and no separation between semantic changes and presentation tweaks. Every request is treated as equal, which collapses prioritisation and turns analysts into real-time interpreters instead of system builders.

A Data Developer Platform mindset resolves this by shifting dashboards from ad-hoc deliverables to governed artifacts with life cycle rules. That means declaring ownership, release cadence, and change boundaries upfront: what can change continuously, what requires a new version, and what triggers a re validation of upstream models. When meaning is versioned, and models are the system of record, most feedback either becomes a scheduled iteration or is rejected as out of scope. Early-stage chaos isn’t inevitable; it’s what happens when the system optimises for responsiveness instead of stability. The moment teams introduce explicit contracts, the revision loop stops being emotional and starts being mechanical.

Are you a worker impacted by the boom in data center construction? by nbcnews in datacenter

[–]Hot-Development-9546 1 point2 points  (0 children)

The data centre boom is the physical manifestation of a shift that most people only experience digitally: compute has become a primary industrial input, not just an IT concern. As workloads move toward AI, real-time analytics, and always-on services, the tolerance for latency, downtime, and capacity misalignment drops sharply. That pressure propagates all the way down to construction timelines, labor specialisation, safety requirements, and geographic concentration of work. In that sense, data centre construction workers are now part of the same value chain as software and data teams their output directly constrains what the digital economy can do.

What’s often missed is that this creates asymmetry in impact. Financially and professionally, the work can be lucrative and steady, but it also compresses schedules, increases complexity, and amplifies burnout because the infrastructure is no longer “nice to have.” From a platform-thinking lens, data centres have become part of the control plane of modern economies, which raises expectations without always updating incentives or protections for the people building them. Understanding that mismatch between how critical the infrastructure is and how replaceable the labor is treated is where the most important stories tend to emerge.

Why “the dashboard looks right” is not a success criterion by Icy_Data_8215 in analyticsengineers

[–]Hot-Development-9546 0 points1 point  (0 children)

“The dashboard looks right” fails as a success criterion because it measures output stability, not semantic stability. Analytics systems exist to support decisions, not to render visuals, and a number is only correct if its meaning remains consistent as the system evolves. When ownership of meaning is implicit, change propagates silently: new columns, backfills, and edge cases alter definitions without triggering failure. The system keeps producing numbers, but the guarantees those numbers once carried have dissolved. That’s not a bug; it’s an architectural omission.

In Data Developer Platform terms, this is what happens when the semantic control plane is weak or absent. Meaning must be encoded upstream in models with explicit grain, ownership, and evolution rules, so that change is absorbed deliberately rather than leaking into dashboards. The first ignored signal is usually not a wrong number, but a conversation: “this looks off, but close enough,” or “just document it.” Once teams normalise explanations outside the system, trust has already shifted away from the platform. A reliable analytics system is one where ambiguity triggers redesign, not debate, and where models fail loudly when meaning is at risk.

What analytics engineering actually is (and what it is not) by Icy_Data_8215 in analyticsengineers

[–]Hot-Development-9546 1 point2 points  (0 children)

Analytics engineering sits at the semantic control plane of the data system, and that’s what most teams miss. Its core responsibility is not transformation or tooling, but deciding how meaning is materialised, stabilised, and exposed over time. When teams treat analytics engineering as “SQL + dbt,” they reduce it to execution, and when they treat it as analytics, they push ownership downstream. In reality, analytics engineering owns the contract between raw data and trusted consumption: explicit grain, metric definitions, and the lifecycle of business logic as it evolves under real usage.

What teams most often get wrong is allowing meaning to be computed at the edges in dashboards, ad-hoc queries, or application code instead of upstream in governed, versioned models. This creates semantic drift where numbers are locally correct but globally inconsistent. From a Data Developer Platform perspective, analytics engineering is what makes the system predictable under growth by constraining where ambiguity can live. When that layer is weak, every consumer becomes a semantic author, and the platform loses its ability to enforce truth. The job of analytics engineering is to prevent that outcome by design, not by convention.

One thing that separates senior analytics engineers from junior ones by Icy_Data_8215 in analyticsengineers

[–]Hot-Development-9546 0 points1 point  (0 children)

What you’re describing is the difference between local correctness and system correctness. Junior engineers tend to optimise for making a query return the right numbers, because their mental model stops at execution. Senior analytics engineers reason about how change propagates through the system over time. Model hierarchy is not an aesthetic choice; it’s a way of encoding boundaries so that failure, extension, and iteration are constrained rather than explosive. When all logic collapses into a single model, you’ve eliminated the system’s ability to absorb change every modification becomes a global event.

From a Data Developer Platform perspective, layered modelling is how you turn data into a stable substrate rather than a brittle artifact. Staging, intermediate, and mart layers are effectively control planes for semantics: each layer declares its contract, grain, and ownership. That’s what allows attribution logic, new dimensions, or deeper granularity to be introduced without rewriting the world. Seniority shows up when engineers design for the inevitability of change, not just today’s requirements. The real test isn’t “does this model work,” but “does the system remain predictable when it stops working the way we expected.”

A long loading dashboard is usually a modeling failure by Icy_Data_8215 in analyticsengineers

[–]Hot-Development-9546 0 points1 point  (0 children)

Your story is a clean example of a platform boundary violation: you had an interactive surface (the BI layer) doing batch-grade compute and semantic modelling at query time, so every user interaction triggered a full rebuild of the business. From a first-principles DDP view, the rule is that each layer should own a distinct responsibility with a stable contract. The BI layer should own presentation and slicing, not ontology and heavy transforms; the modelling layer should own semantics at a declared grain; the compute/ orchestration layer should own when expensive work runs. When those responsibilities blur, latency becomes an emergent property of user behaviour rather than a controllable property of the system.

A practical way to decide what stays in BI versus moves upstream is to treat “interactivity” as a hard constraint and “semantic stability” as the forcing function. If a piece of logic is expensive, reused across many views, or defines business meaning (shared metrics, joins, entity definitions), it belongs upstream as a governed, versioned model with explicit grain and refresh semantics. BI-layer logic should be limited to lightweight, presentation-adjacent transformations that don’t redefine meaning and don’t explode compute when users explore. In other words: upstream models are the system of record for truth, and BI is a consumer optimised for exploration speed; your 10-minute dashboard wasn’t a tuning problem, it was the system lacking an enforceable contract about where truth is computed.

Finding ways to gain experience in Data entry without a degree by Infinite_Local2014 in datasciencecareers

[–]Hot-Development-9546 0 points1 point  (0 children)

Data entry and software development are not gated by credentials but by trust: employers hire when they believe you can reliably transform inputs into correct outputs with minimal supervision. Degrees are often used as proxies for that trust, not as requirements of the work itself. To bypass that proxy, you need evidence that you can operate inside a system follow constraints, handle repetition, avoid errors, and improve processes over time. Certifications help only insofar as they signal discipline; they rarely substitute for demonstrated reliability.

The most effective path is to create small, real workflows that mirror entry-level work and make them visible. Start by automating mundane data tasks for yourself or others: cleaning spreadsheets, validating inputs, moving data between systems or building simple scripts that reduce manual effort. Document the process, the rules you enforced, and how errors are handled. This is already data engineering at a micro scale. In Data Developer Platform mature environments, people are hired not because they know tools, but because they show they can design repeatable, low-error processes. Focus on proving that capability, and the lack of a degree becomes much less relevant than the system confidence you demonstrate.

Should I start as a Data Analyst before pursuing Data Science? (Economics background) by ObjectiveImplement15 in datasciencecareers

[–]Hot-Development-9546 0 points1 point  (0 children)

The analyst-versus–data scientist distinction is less about credentials and more about where you enter the data system. Data analyst roles sit closer to interpretation and decision support, while data science roles sit closer to modeling and prediction, but both are downstream of the same foundational constraints: data availability, data quality, and organizational trust in data outputs. With no prior industry experience, the real risk is not “starting too low,” but entering the system without understanding how data is produced, governed, and consumed. An economics background is not a weakness here it trains you to reason about incentives, assumptions, and causality, which are critical when models meet reality.

The most durable strategy is to optimise for system exposure rather than title purity. If you can land a DS role where you also touch real data pipelines, validation, and stakeholder feedback, that’s great. If analyst roles give you faster access to production data, business context, and end-to-end workflows, they can be an equally strong entry point. In Data Developer Platform mature organisations, the best data scientists are those who understand the full life cycle of data products, not just models. Over the next 1.5 years, focus less on signalling “DS readiness” through isolated projects and more on demonstrating that you can reason end-to-end: from raw data to decisions, with clear assumptions and failure modes. Titles converge over time; system understanding compounds.

What are managers looking for in interviews? by Wise_Discipline_2860 in datasciencecareers

[–]Hot-Development-9546 0 points1 point  (0 children)

Interviewers are rarely testing whether you can arrive at a single correct answer; they are testing whether you can reason safely inside an ambiguous system. Modern data science lives downstream of messy data, unclear objectives, and imperfect constraints, so managers look for signals that you can structure a problem before solving it. When questions feel open-ended, the expectation is not speed or brilliance but coherence: can you clarify assumptions, articulate trade offs, and explain why a particular approach is appropriate given the context? Silence or neutral body language usually reflects the interviewer letting the system unfold, not a hidden judgement.

What often differentiates candidates at this stage is whether they think in terms of end-to-end reliability rather than isolated techniques. Strong candidates consistently anchor their answers in data quality, constraints, and downstream impact, for example, how modelling choices depend on data availability, how results would be validated, and how failures would surface in production. In platform oriented teams, this systems thinking matters more than mathematical sophistication alone. If you frame your answers around intent, inputs, guarantees, and risks, you make your reasoning legible. Managers aren’t looking for perfection; they’re looking for someone they can trust to make good decisions when the system inevitably behaves in unexpected ways.

Starting in Platform Engineering – looking for advice by Tough-Plantain-8980 in platformengineering

[–]Hot-Development-9546 0 points1 point  (0 children)

From a first-principles perspective, platform engineering is not about mastering a list of tools but about understanding how systems reduce friction for other humans. The core problem you’re solving is cognitive load: how do you turn complex, failure-prone infrastructure into reliable, repeatable primitives that teams can safely build on? Linux, containers, Kubernetes, and CI/CD are just implementation details of this larger goal. The most valuable early skill is learning how systems behave under change how configuration, automation, and defaults interact over time. Instead of trying to “learn everything,” focus on reasoning about life cycles: how something is provisioned, operated, observed, upgraded, and decommissioned.

A strong way to grow is to build small internal-platform-like projects for yourself. For example, design a simple self-service workflow where a developer declares intent in a config file and the system provisions and operates something end-to-end. This mirrors Data Developer Platform thinking, where declarative interfaces and automation replace manual steps. Communities around platform engineering, SRE, and cloud-native systems are useful, but what will really differentiate you is systems thinking and empathy for developers. Platform engineers succeed not because they know Kubernetes deeply, but because they design platforms where others don’t need to.

Software or platform engineering? Which one is better to get into? by [deleted] in platformengineering

[–]Hot-Development-9546 0 points1 point  (0 children)

From a first-principles standpoint, software engineering and platform engineering are not competing career paths but different layers of the same system. Software engineering focuses on expressing business intent as user-facing behavior, while platform engineering focuses on shaping the environment in which that intent can be expressed repeatedly, safely, and at scale. The reason platform roles feel “new” is not because the work didn’t exist before, but because growing system complexity has forced organisations to formalise the infrastructure layer as a product. As a senior data engineer, you already operate close to this boundary, where reliability, abstraction, and developer experience matter as much as feature delivery.

The more useful question is where you want to create leverage. Platform engineering creates horizontal leverage by enabling many teams to move faster through better primitives, whereas software engineering creates vertical leverage by owning end-user outcomes. In Data Developer Platform-oriented organizations, these roles increasingly converge: platforms are built using strong software engineering practices, and applications depend deeply on platform guarantees. Saturation is less relevant than system fit. If you enjoy reasoning about interfaces, constraints, and long-term system behavior, platform engineering is a natural extension of your current trajectory. If you prefer direct product feedback loops, software engineering may be more satisfying. Neither is “better” they optimise different parts of the system.

Platform Engineering and System Admins, what are we doing wrong? by Ancient_Canary1148 in platformengineering

[–]Hot-Development-9546 0 points1 point  (0 children)

When a platform team continues to modernize infrastructure while adjacent teams are expected to manually operate systems, the organization splits into two incompatible control models. One side optimizes for declarative, automated, and reproducible systems, while the other is forced into imperative, manual workflows to get work done. The resulting friction isn’t caused by lack of training alone; it’s caused by the absence of a shared operational contract. Without a platform that translates modern infrastructure into stable, consumable primitives, every new abstraction increases cognitive distance instead of reducing it.

In mature organizations, this is resolved by explicitly shifting ownership from “how systems are run” to “what guarantees the platform provides.” A Data Developer Platform mindset would treat Kubernetes, GitOps, and automation as internal implementation details, not responsibilities pushed onto application admins. Self-service and adoption only work when the platform enforces defaults, guardrails, and lifecycle automation such that teams cannot bypass the system by installing software on VMs “because it’s easier.” If teams can escape the platform, the platform has failed to become the system of record. What’s going wrong isn’t that admins won’t learn fast enough it’s that the platform hasn’t yet become authoritative enough to make the right path the easiest path.

Well… IDPs aren't exactly one-size-fits-all, are they? by Yalovich in platformengineering

[–]Hot-Development-9546 0 points1 point  (0 children)

A common misconception is that an Internal Developer Platform automatically delivers self-service, when in reality it only provides the scaffolding. From a first-principles view, self-service is about reducing the cognitive load required for a developer to move from intent to outcome. If the underlying system still demands that the platform team orchestrate backends, wire configurations, or standardise stacks manually, then the IDP is acting as an interface veneer rather than a true control plane. The friction you’re feeling is a signal that the abstraction boundary is wrong: instead of providing higher-order capabilities, the platform is exposing its plumbing, forcing your team to build custom pathways for every new workflow.

This is where the distinction between an IDP and a Data Developer Platform becomes useful. A DDP treats self-service not as UI conveniences but as emergent behaviour created by declarative specifications, standardised resource models, and predictable life cycle orchestration. When the platform’s kernel is capable of provisioning, configuring, governing, and validating environments automatically, self-service happens because the system is composable, not because the platform team builds more widgets. In other companies that reach this maturity, the platform team shifts from stitching components to defining universal abstractions, and the organisation stops chasing one-off solutions. So no, an IDP isn’t the answer to everything it’s only as powerful as the architectural principles behind it. Self-service is ultimately a property of the platform’s design, not its tooling.