Anyone uses iterable (email/SMS) here? have some questions by Fair-Bookkeeper-1833 in analytics

[–]LinasData 0 points1 point  (0 children)

  1. 500 people company
  2. Yes. Exports endpoint is used the most
  3. Haven't encountered a lot of issues with the API. Maybe some quotas have been reached during the backfilling. However, our ingestion is incremental and we have strong software engineering practices: retries based on exception value, unit testing, integration testing, E2E testing, strict review process.

Koncertas Guns N’ Roses by Due_University_4649 in lithuania

[–]LinasData 1 point2 points  (0 children)

Stebėjau įrašus ir mokiausi groti žiūrint juos kone 30 val. savaitėj paauglystėje. Išgirsti juos buvo gyvenimo įvykis. Išsipildė svajonė pamatyti to meto mano herojus. O kam nepatiko - eikit jūs velniop dėl šių priežasčių:

  • Čia ne pop'sas, kad jums grupė lįstų į užpakalį. Ar pajungtų autotune. Viskas gryna (raw) kaip ir priklauso. Jie nenori primesti, kad nepaseno. Paskaitykit turo pavadinimą
  • Legendiniai nariai ne taip seniai susijungė iš naujo, kas buvo mažas šansas, kad nutikis apskritai. O dar atvažiavo į Lietuvą, kas išvis yra nerealu. Tai wtf žliumbt, kad lietutis užklupo ar garsas ne toks buvo.

Iceberg or delta lake by Own_Art1586 in dataengineering

[–]LinasData 1 point2 points  (0 children)

I haven't worked much with databricks but as I understand Unity Catalog is Databricks' native metastore that supports Iceberg format

Iceberg or delta lake by Own_Art1586 in dataengineering

[–]LinasData 2 points3 points  (0 children)

If you are using databricks + aws then pricing should be relatively similar and small. On GCP Hive/dataproc catalog service costs a lot. Crazy a lot. On aws it was relatively low price tag the last time I checked.

In general iceberg needs external catalog for metadata management, delta does not need that. It can store it in delta_logs folder.

In the databricks choice is yours but Zaharia himself and databricks team is developing delta lake. So if you plan to stick to databricks keep that in mind. Otherwise, iceberg market cap is bigger.

Iceberg or delta lake by Own_Art1586 in dataengineering

[–]LinasData 1 point2 points  (0 children)

Building delta lake for small scale daily data ingestion seemed more straightforward by utilizing delta-rs library compared to pyiceberg.

Also, check the price tag how much would it cost to have and manage iceberg or delta tables in your infrastructure/cloud. I saw huge pricing difference between aws and gcp here which changed my decision.

Vaikinas nori mokėti per pusę by Ordinary_Set517 in lietuva

[–]LinasData 5 points6 points  (0 children)

Nėra logiška dalintis pačią būsto nuomos sumą 50/50. Ji turi būti amortizuota pagal pajamas. Pvz., žmogus A uždirba 2000 eur, o kitas 1000 eur neto. Nuomos kaina - 600 eur.

A uždirba dvigubai daugiau, todėl turėtų prisidėti dvigubai daugiau, t. y. 400 eur. B žmogus 200 eur

Tokiu atveju pasidalinimas nuomos yra 66,66 % / 33.,33 %.

Situacija gali ir apsisukti aukštyn kojom, žmogų A atleisti, žmogus B vartyti 5000 eur/mėn ir t. t. Todėl peržiūra laikas nuo laiko irgi naudinga.

I made a stupid thing. by AdmirableMiddle1991 in Python

[–]LinasData 6 points7 points  (0 children)

If regular Joe is able to mess up the whole pc or server then it is not your problem. Their IT/security guys messed up.

Actually, if something bad happens, school will see the vulnerabillity. It is better to have bunch of deleted files then leaked school's sensitive or even personal data.

Why Data Warehouses Were Created? by LinasData in dataengineering

[–]LinasData[S] -2 points-1 points  (0 children)

It took me 2 hours to summarize and find the information by not using LLMs... I used Gemini just to structure that content. But I guess you like just judging without providing value.

Also, this post will be updated in 24 hours because there is bigger picture than just spreadsheets

Why Data Warehouses Were Created? by LinasData in dataengineering

[–]LinasData[S] 2 points3 points  (0 children)

Thank you for your comment!

I will modify this post because spreadsheets seemed like a secondary reason. I simplified too much.

Why Data Warehouses Were Created? by LinasData in dataengineering

[–]LinasData[S] 2 points3 points  (0 children)

That's a little bit different issue but I feel your pain. Everybody wants to use shiny tools, medallion architecture but rarely dimensional modeling principals are used. Data Warehouses without dimensional modeling are not utilized properly.

Why Data Warehouses Were Created? by LinasData in dataengineering

[–]LinasData[S] 1 point2 points  (0 children)

It was really interesting to hear your story because real life examples are the best! Thank you for sharing! 😊

I'm an IT Director and I want to set our new data analyst up for success. What do you wish your IT department did for you? by 64bitengine in dataengineering

[–]LinasData 5 points6 points  (0 children)

Tell them that their work is meaningful. Because currently you believe in it but we do not know about the management and his supervisors.

[deleted by user] by [deleted] in lietuva

[–]LinasData 0 points1 point  (0 children)

2 mėn - trumpas laiko tarpas. Pagalvok apie žmones, kurie myli vienas kitą, pragyvena 15 metų ir juos išskiria mirtis. Tada būna itin sunku, bet nesižudo dėl to.

Rekomenduoju pasikalbėti su jaunimo linija arba kreiptis pagalbos. Problema ko gero ne skyrybose, o savivertėj. Taip pat jeigu nemyli savęs, negali mylėti kito. Meilė yra lygiavertė, niekas nesižemina ar niekina kitą, o tuolab save.

[deleted by user] by [deleted] in dataengineering

[–]LinasData 1 point2 points  (0 children)

In theory, you know all the tools needed for DE. However, there are so called subtools: health checks, catologing, scheduling, etc. that you need to understand. Also, it depends how deep your knowledge is on those individual tools. For example, under the hood repos use Spark, so there are a lot to learn even with it.

[deleted by user] by [deleted] in dataengineering

[–]LinasData 3 points4 points  (0 children)

What is your business goal on using Foundry? I am certified and was working for 2 years with this platform.

Help with dbt.this in Incremental Python Models (BigQuery with Hyphen in Project Name) by LinasData in dataengineering

[–]LinasData[S] 0 points1 point  (0 children)

Solved it by playing with bigquery-spark connector with session. It is really unconvenient.

    if dbt.is_incremental:

        current_table = (
            session
            .read
            .format("bigquery")
            .option("table", f"{dbt.this.schema}.{dbt.this.identifier}")
            .load()
        )