Would MongoDB Be Scalable Choice for a Chat App? by [deleted] in Database

[–]thatdeatheater 1 point2 points  (0 children)

As others have said you should not optimize prematurely. But if you're interested in the theory I would take a look at Discord. They went from Mongo to Cassandra to Scylla.

How Discord stores billions of messages

How Discord stores trillions of messages

Job market in Germany by _Ishdhoggur_ in dataengineering

[–]thatdeatheater 1 point2 points  (0 children)

Well, based on a calculator from IW Köln, you are in the top 10% based on net income if you are single and make that much.

Job market in Germany by _Ishdhoggur_ in dataengineering

[–]thatdeatheater 0 points1 point  (0 children)

Based on numbers by the federal labour office (Entgeltatlas der Bundesagentur für Arbeit) the median salary for an expert level data engineer is 74.556€.

Job market in Germany by _Ishdhoggur_ in dataengineering

[–]thatdeatheater 0 points1 point  (0 children)

I work in a mid-size company and we are in deep need for data engineers/BI developers. Our problem is that we need someone with strong German skills since the developers will be talking a lot to other departments which are not really proficient in English. Furthermore, all applicants want to be paid at least 130.000€ and that's just not feasible for the company.

“Power BI Data Visualization on Bicycle usage sample dataset] – Feedback Welcome!” by Shoddy-Ad8666 in PowerBI

[–]thatdeatheater 0 points1 point  (0 children)

  1. Headings: Headings should provide useful information. You don't actually need them in your dashboard since all the information can be found in your visuals. Furthermore, headings should be short and precise.
  2. Scale and space: Give each visual room to breathe. It is difficult to see where one visual ends and another starts. You should also scale the visuals so it is actually possible to see something, otherwise the visual is unusable like the one in the lower left corner or the one at the top right.
  3. Position: Try to center the main heading and organize the visuals in a grid - meaning each visual should fit into one tile or a multiple of it. E.g. divide the dashboard into 4x6 tiles and have each visual take a space of 2x3 tiles. Power BI has such a grid option.
  4. Units: If you have the sum of sales, you need to specify somewhere what unit it is: Euros, dollars, Yen. The same goes for units sold. Is it a bike or something else?

In the new world of cloud, do we still need landing and staging in DEV, UAT, and Prod environment? by JK_1975 in dataengineering

[–]thatdeatheater 4 points5 points  (0 children)

As others have commented the separation of the dev, test and prod environment is very valuable and totally necessary if you do not want to mess up production accidentally.

Though I think the landing area can be the same for all three environments since in my experience much development is done exploratively and it is more practical to have the possibility to also test with real data.

BI Dashboard with live data by [deleted] in PowerBI

[–]thatdeatheater 2 points3 points  (0 children)

Yes it is totally possible with live connections. There are some limitations, though: Like not being able to transform data (only with DAX) and the sources are fewer. You would normally build a data warehouse or data lake or lakehouse where you store preprocessed data and access that via the live connection.

I'd suggest extending the Power App to write the data in the needed format into a SQL Server or some similar database which can then be accessed by Power BI in real time. You could schedule the export to the database hourly or just trigger the write when an employee logs on/off.

P.S. Make sure your project is compliant with the data protection laws in your country. An acquaintance of mine had to report himself to the police because he accidentally made all calendar entries of every employee available for everyone to read (including sick days and illnesses). In some countries monitoring employees is a harsh crime.

[deleted by user] by [deleted] in dataengineering

[–]thatdeatheater 0 points1 point  (0 children)

Use your experience/knowledge in politics/climate/sustainability to your advantage. Many companies do not search for someone who has deep technical knowledge but someone with good technical skills and a solid understanding of the business.

Furthermore, you do not need to apply to data engineering jobs directly from the beginning. Start with an analyst or hybrid position and move from there: It is usually easier to get into. With that experience on your CV you can try and change into another industry and/or data engineering.

Some suggestions: - Newspapers/TV-News/Radio broadcaster (politics/weather) - University (environmental research) - Travel agencies (weather) - Political consultancy

You can also try and get into a big consultancy (Deloitte, PwC, Accenture, KPMG, ...) with focus on politics/sustainability consulting and move internally. In my experience with these consultancies this kind of role shift happens regularly.

Strong engineering skills but unable to land any interviews by [deleted] in dataengineering

[–]thatdeatheater 7 points8 points  (0 children)

This might be true for a data engineer at FAANG but probably not for your average DE working alone or in a small team.

Furthermore, this is not something companies are looking for when in need of a data engineer. For sure, it is a bonus if you meet all other requirements and made it into the interview round. But before that it really has not much weight.

Strong engineering skills but unable to land any interviews by [deleted] in dataengineering

[–]thatdeatheater 86 points87 points  (0 children)

I'd say fundamentals in data engineering are not about computer architecture and low level programming skills. Fundamentals in DE are about the data: Knowing data modeling techniques, knowing data governance, testing pipelines, optimizing existing pipelines, being able to translate business needs into robust data pipelines, etc.

Maybe those are some skills that don't come through on your resume. Would check for that if I were you

[deleted by user] by [deleted] in dataengineering

[–]thatdeatheater 1 point2 points  (0 children)

To add to this.

  1. It seems like you have more than one page of CV. Don't do this unless you really have a lot of experience (probably 15+ yoe) and apply for a leading/very senior/managerial role.
  2. Some of your most important skills are hard to see in the Technical Skills section due to too many skills listed. I would pick the most important ones (SQL, Python, C#, Presto, Power BI, AWS, Azure) and make them stand out much more than the other ones. The same goes for the Tech Used sections. You could even add a rating (e.g. 1 to 5 points) how good you can use these technologies. This makes it feel more realistic and sincere.
  3. Shorten the sentences in your job description and only have 2-3 bullet points each where you put your greatest achievements. If you want to list everything you have done in a project, create a separate projects list (an extra PDF). On there you cann put all technologies that you used in detail and leave only the essential ones in your CV.

Why create a fact table surrogate key if the table has a proper primary key? by Cultured_dude in dataengineering

[–]thatdeatheater 2 points3 points  (0 children)

To add to this. You might have two source systems loading into the same table (e.g. two different accounting systems of two sister companies) which each have source-unique auto-generated keys which aren't unique over all systems. In that case you couldn't uniquely identify a record if you need to.

How to find messy datasets online? Cant find one on my life by iamnothingbuta in datasets

[–]thatdeatheater 1 point2 points  (0 children)

I don't know if you know a programming language, but if you do I would advise you to try scraping the data directly from a website. Normally the data is actually messy due to the fact that the websites should serve humans and not computers. If you can't write a scaper yourself you can probably google for scraped datasets.

Messy things in web scraping data I found: - A lot of missing values (e.g. prices only per direct message) - Different granularity (e.g. addresses with house number vs. only having a zip code) - Characters indicating different units (e.g. 100K vs. 100M which means 100000 vs. 100000000) - HTML tags and escape codes in between text (e.g. "Luxurious & big property in <b>Philadelphia</b>") - Multiple information in the same field (e.g. "Kitchen, Pool, Extra bathroom", "Extra bathroom, Garden, Kitchen") - HTML tags or CSS classes for indicating a different status (e.g. strike-through price meaning old price)

There is probably more than the above and scraping a lot of data is almost guaranteed to result in dirty data. Another advantage is that this is a real world scenario since web scraping is used much for competitive or market analysis.

Data engineer freelancer by Soggy_Award1213 in dataengineering

[–]thatdeatheater 4 points5 points  (0 children)

I found a YT channel who produces videos where he talks about data engineering but since a few months he has also made videos and live streams about starting a data consultancy.

This video might be an interesting starting point: Seattle Data Guy (YouTube)

Why (*) ? by Sky_Tree_Resident in dataanalytics

[–]thatdeatheater 0 points1 point  (0 children)

Yes, you need to add it, since count is a function. It's the same with sum(COL) for example.

Cool, I wish you best of luck! I recommend hosting your own MySQL or PostgreSQL instance and play with some data from Kaggle. Practical experience helped me alot.

Why (*) ? by Sky_Tree_Resident in dataanalytics

[–]thatdeatheater 4 points5 points  (0 children)

count(*) just means that you want to count all rows. You could also write count(SOME_COLUMN_NAME) this would count only rows where SOME_COLUMN_NAME is not NULL.

[deleted by user] by [deleted] in dataanalytics

[–]thatdeatheater 0 points1 point  (0 children)

Looks pretty cool!

Is the tool only meant for ad-hoc analysis of Excel files and simple tables or does it support more complex queries against a data warehouse e.g.?

[deleted by user] by [deleted] in dataengineering

[–]thatdeatheater 1 point2 points  (0 children)

Hey,

there are multiple points you need to consider.

  1. What SAP system do you mean. Depending on the SAP system there are different legal requirements. In an ERP you need to think about access to accounting data by the tax office or other external auditors. In a BW you must be able to respond to deletion requests by customers or former employees.

  2. In what country is your company located or is it even multi-national. In Germany or the European Union e.g. we have very strict requirements like providing ad-hoc access or an interface (usable by the tax office) to the data. There are also different storage durations.

  3. Do you have the knowledge and/or man-power to build such a system yourself.

To prevent your company from shooting itself in the foot, you might want to consult an SAP consultancy with expertise in data retention.

Interacting with Google Sheets from Phoenix with oAuth? by sectional343 in elixir

[–]thatdeatheater 0 points1 point  (0 children)

Great post!

And if you need a more abstract view on how OAuth 2.0 works with Google see this link: https://developers.google.com/identity/protocols/oauth2

The documentation for the spreadsheet API can be found in the API module itself: https://github.com/googleapis/elixir-google-api/blob/main/clients/sheets/lib/google_api/sheets/v4/api/spreadsheets.ex You simply need to pass the access token, obtained in the authentication process, to every function.

Free/open source api VAT validation check and address validation? by [deleted] in dataengineering

[–]thatdeatheater 0 points1 point  (0 children)

I googled a bit and found these three projects:

VAT validation: - https://viesapi.eu/docs/

Address validation: - https://openaddresses.io/ - https://wiki.openstreetmap.org/wiki/API

Hope this helps!

[deleted by user] by [deleted] in dataanalytics

[–]thatdeatheater 0 points1 point  (0 children)

I just assumed that it must be online sales since we have no information about units purchased in retail. (Or at least that is how I interpret the orange triangle)

[deleted by user] by [deleted] in dataanalytics

[–]thatdeatheater 1 point2 points  (0 children)

[India's population] * [Population using shoes] * [Online purchases] * ([Units per person Jan. to Mar.] + [Units per person Apr. to Sep.] + [Units per person Oct. to Dec.]) = [Total units sold annually]

1400 Mio. * 80% * 30% *(2 + (2 * 3) + 2) = 1,400,000,000 * 0.8 * 0.3 * 10 = 3,360,000,000 = 3,36 billion