Advancing Spark - Give your Delta Lake a boost with Z-Ordering

แชร์
ฝัง
  • เผยแพร่เมื่อ 1 ต.ค. 2024

ความคิดเห็น • 24

  • @nadezhdapandelieva3387
    @nadezhdapandelieva3387 ปีที่แล้ว +7

    Hi Simon, I like your videos, they are super useful. Can you make some videos on how to optimize jobs and reduce the performance time or how to investigate when optimization is needed on the job?

  • @Reader-ju1yn
    @Reader-ju1yn 2 หลายเดือนก่อน

    Super detail and clean explanation. Love this z-ordering concept.

  • @sarmachavali7676
    @sarmachavali7676 2 ปีที่แล้ว +1

    Hi Simon, Nice video and is useful. I have a quick question, we are replicating huge data from MSSQL Datawarehouse to Deltalake using DLT(including CDC changes) with continuous mode .As part of that, i have specified my zorder is same as primary key; Does this increases the performance of merge operation in (apply statement) or not.How can i check this performance metrics.

  • @nayeemuddinmoinuddin2186
    @nayeemuddinmoinuddin2186 2 ปีที่แล้ว +1

    Hi Simon - Thanks for this awesome video. One quick question , Do Optimize and Z-order disturb the checkpoint in case of Structured Streaming?

  • @dmitryanoshin8004
    @dmitryanoshin8004 3 ปีที่แล้ว +1

    Can I have partition by date and Zorder by event name? Or partition and Z should be same columns?

  • @PersonOfBook
    @PersonOfBook 3 ปีที่แล้ว +1

    Can you use both partition by and zorder by, on the same column or different columns. And if so, would it be beneficial? Also, why do you enclose spark.read with brackets?

    • @AdvancingAnalytics
      @AdvancingAnalytics  3 ปีที่แล้ว +4

      Hey - so you /can/ z-order by a column you've partitioned on, but it'll give no benefit as your data is already sorted into those values by the partitioning!
      And brackets around the spark statement means you can span multiple lines without needing a line escape '\' for every line!

  • @vt1454
    @vt1454 ปีที่แล้ว

    Great Videos Simon. One suggestion on background ribbons of slides. The ribbons on your slide templates keep moving and are bit uncomfortable to eyes. Request if this can be static

  • @dheerajmuttreja
    @dheerajmuttreja ปีที่แล้ว

    Hey Simon .. Great explanation with proper use snd demo

  • @kingslyroche
    @kingslyroche 4 ปีที่แล้ว +1

    good explanation! thanks.

  • @the.activist.nightingale
    @the.activist.nightingale 4 ปีที่แล้ว +1

    Simon is back!!!!
    Thank you for this awesome video :) Could you make one explaining how we can profile a spark script in order to identify optimizing tuning opportunities ? I always fo the the Spark UI but I'm completely lost. I know one thing for sure, too much is swapping between nodes is bad news :)!

    • @AdvancingAnalytics
      @AdvancingAnalytics  4 ปีที่แล้ว +5

      Oooh, ok, so a quick tour of the Spark UI and "some things to look out for" when diagnosing spark performance problems? I'll add it to the list - need to thing about what the top ones would be or it'll be two hours long!
      Simon

    • @the.activist.nightingale
      @the.activist.nightingale 4 ปีที่แล้ว

      Advancing Analytics You’re the real MVP Simon! TY!!

  • @DebayanKar7
    @DebayanKar7 ปีที่แล้ว

    Awesome-ly Explained !!!!

  • @cchalc-db
    @cchalc-db 3 ปีที่แล้ว

    Can you share the NYTaxi notebook?

  • @preethi7674
    @preethi7674 2 ปีที่แล้ว

    In production environments, do we have to zorder the tables weekly to improve performance?

    • @workwithdata6659
      @workwithdata6659 8 หลายเดือนก่อน

      Yes. You will have to z order on regular basis. And there is no guarantee that only new files will be re-written. Running optimize on big tables which get good size of incremental data can be counter productive.

  • @AndreasBergstedt
    @AndreasBergstedt 4 ปีที่แล้ว +1

    1st :)

  • @ipshi1234
    @ipshi1234 4 ปีที่แล้ว

    Thanks Simon for the great video! I'm curious if I would have to achieve Z-ordering in Delta Lake Synapse, how would I be able to? As the Optimize command is only available on the Databricks runtime? Thank you :)

    • @AdvancingAnalytics
      @AdvancingAnalytics  4 ปีที่แล้ว +2

      Hey!
      On the file optimisation level, you could maybe achieve something similar using bucketing - but you wouldn't get the same data skipping benefits. Probably easier to just spin up a databricks cluster over the same data and use that for maintenance jobs (again, Synapse wouldn't do the data skipping part, but your files would be arranged properly)
      For the indexing/query performance side - Microsoft have been building "Hyperspace", which is an indexing system separate to Delta. This might be the answer for where you can't optimize tables...but it's a very early product, I've not had a go at using it yet!
      Simon

  • @devanssshhh
    @devanssshhh 3 ปีที่แล้ว

    hey Thanks its a great video.

  • @katetuzov9745
    @katetuzov9745 ปีที่แล้ว

    Brilliant explanation, well done!

  • @vishalaaa1
    @vishalaaa1 ปีที่แล้ว

    excellent

  • @nsrchndshkh
    @nsrchndshkh 3 ปีที่แล้ว

    Thank you very much