Spark Interview Question | Scenario Based Question | Explode and Posexplode in Spark | LearntoSpark

แชร์
ฝัง
  • เผยแพร่เมื่อ 26 ธ.ค. 2024

ความคิดเห็น • 21

  • @SatishKumar-fr9yv
    @SatishKumar-fr9yv 2 ปีที่แล้ว +3

    Small correction: explode() won't take Struct types; It takes only array and map types

  • @sravankumar1767
    @sravankumar1767 2 ปีที่แล้ว

    Superb explanation 👌

  • @localmartian9047
    @localmartian9047 2 ปีที่แล้ว +1

    I checked and posexplode is still not working with withColumn in spark 3. Analysis exception

  • @adityapratapsingh7649
    @adityapratapsingh7649 3 ปีที่แล้ว

    Difficult question. Easy solution 👍

  • @rajdeepsinghborana2409
    @rajdeepsinghborana2409 3 ปีที่แล้ว +1

    Nice

  • @usharani-kw6gt
    @usharani-kw6gt 4 ปีที่แล้ว

    Nice explanation bro

  • @yaniv54
    @yaniv54 4 ปีที่แล้ว

    Thank you for the series, it is very useful

  • @anil099kumar6
    @anil099kumar6 3 ปีที่แล้ว

    Hi bro in pyspark dataframe how to find each column null values count

  • @sumantaghosh9299
    @sumantaghosh9299 3 ปีที่แล้ว

    Hi Shahul , in this example we have only one array , now we have multiple struct array . can you pls explain how to do that ? thanks in advance mate

  • @SpiritOfIndiaaa
    @SpiritOfIndiaaa 4 ปีที่แล้ว

    Thanks , can you please make some samples using agg with struct of arrays??

  • @muraliintooru1468
    @muraliintooru1468 3 ปีที่แล้ว

    Could you please create a video on "how to implement SCD type 2 using pyspark"?

  • @rohithreddysureddy9151
    @rohithreddysureddy9151 2 ปีที่แล้ว +1

    What if there are two or more columns that need to be exploded ??

    • @AzarudeenShahul
      @AzarudeenShahul  ปีที่แล้ว

      We cannot use more than one explode in a spark sql statement, we have work around for this, will discuss in upcoming videos..
      Thanks for your support 🙂

  • @chaitanyakrishna5873
    @chaitanyakrishna5873 3 ปีที่แล้ว +1

    can you post all possible interview questions

    • @AzarudeenShahul
      @AzarudeenShahul  3 ปีที่แล้ว

      Sure, Please have a look at our Playlist, we have 40+ interview scenario question. We will add more :). Please do try on your own machine for better understanding.

  • @sangamrathore7850
    @sangamrathore7850 3 ปีที่แล้ว

    Hey Azar,
    I need your help here.
    I am trying this example but getting this issue below
    This is the input file I am using
    input file - explode_posexplode_ex.csv
    Name|Age|Education
    Azar|25|MBA,BE,HSC
    Sangam|27|BE,HSC
    Dilip|MBA,BBA,HSC
    Ajay|BE,HSC
    then I created df as below
    df = spark.read.option("delimeter","|").csv("/FileStore/tables/explode_posexplode_ex.csv",header=True)
    But I am getting only one col instead of 3. When i execute below command output is shown below
    df.show()
    df.columns
    (1) Spark Jobs
    +------------------+
    |Name|Age|Education|
    +------------------+
    | Azar|25|MBA|
    | Sangam|27|BE|
    | Dilip|MBA|
    | Ajay|BE|
    +------------------+
    Out[42]: ['Name|Age|Education']

  • @maheshtej2103
    @maheshtej2103 2 ปีที่แล้ว +1

    bro how to remove delimeter from spark datafrme?????? i need answer brooo

    • @AzarudeenShahul
      @AzarudeenShahul  2 ปีที่แล้ว

      Can u share some sample inp and out here or in mail

  • @srinivasasameer9615
    @srinivasasameer9615 4 ปีที่แล้ว

    How to do with Scala please help me

    • @AzarudeenShahul
      @AzarudeenShahul  4 ปีที่แล้ว

      Please mail me the error that you are facing