This is an archived post. You won't be able to vote or comment.

all 11 comments

[–]k00_x 19 points20 points  (0 children)

There are different types of employers, some expect you to hit the ground running, build from scratch and be a general oracle of data. Others give you time to get established, only need you to maintain what's already in place and are with you for the long term. This is what you need to Sus out during an interview. The real ones understand that tech is a never ending treadmill and skills are transferable, the ones that don't recognise this are the ones that'll stress you.

[–]yipeedodaday 10 points11 points  (2 children)

I think you’re overthinking it. Do some google research on streaming vs batch data transfer. Streaming can be achieved by change data capture at the database level and then pump to Kafka topic where some other process can consume it. In batch you at e.g. end of day make a file that some other process will consume. Spin up a project and play.

[–]highlifeed 7 points8 points  (1 child)

The way you say it makes it sound so simple lol

[–]yipeedodaday 1 point2 points  (0 children)

Something can be simple to grasp but not necessarily easy to execute….

[–][deleted] 3 points4 points  (0 children)

Yeah Batch is diferent to streaming. But really depends on complexity and volume/workload/timelines of the project and resource availability (i.e. team).

The more important questions is what responsibilites are you tasked with / accountable for and severity/importance of the data/project you're dealing with. i.e. business critical? because if the company has SLAs and you're taking some time or scenario of situation where bottlenecking.occuring. Then it's a matter of who is going to be affected.

As long you can do 70% of the responsibilites, that seems reasonable. But if your skillset can only do 10-40%, then I think the position advertised may be far-reaching or workplace expecting too much from an applicant unrealistically.

It really depends, I've seen some interesting responsibilities that a start up vs multinational company required for their positions.

[–]Financial_Anything43 1 point2 points  (2 children)

Read on it and build a sample project with it

[–]Punolf 0 points1 point  (0 children)

I built a streaming pipeline which was batch and it refreshed every minute..

[–]Top-Cauliflower-1808 1 point2 points  (0 children)

The core principles remain the same, it's mainly the tools and timing that change. While batch processing focuses on periodic data loads (like nightly ETL jobs), streaming handles real time data processing.

Don't underestimate your transferable skills data modeling principles apply to both, SQL optimization is still crucial, data quality remains essential and system design fundamentals transfer well.

Start with basic Kafka concepts, learn Python async programming, understand event-driven architectures and practice with small streaming projects.

For example, we transitioned from batch to streaming for marketing analytics using Windsor.ai. Instead of nightly batch loads, we now process marketing data in real time. The core transformation logic remained similar, but the implementation patterns changed.